Mason researchers examine computational biology approach to flow diversion

Juan Cebral, George Mason University Professor of Bioengineering, and his collaborators are constructing computational fluid dynamics models of cerebral aneurysms to compare the effects of flow diverters and assess the hemodynamic differences between immediate occlusions and long-term patency.

To do so, they are using 3-D rotational angiography images and data on the effects of flow-diverting devices.

They will include both flow-diverting stents as well as intra-saccular devices in their simulations.

Cebral and his collaborators are using previously developed methods based on unstructured grids, embedding and immersed techniques to model flow diverting devices with high-resolution adaptive meshes.

In addition, Cebral and his team are creating sophisticated models of the brain circulation and using them to investigate possible causes of hemorrhages observed after treatment of cerebral aneurysms with flow diverting stents. {module INSIDE STORY}

The researchers will conduct numerical simulations using software developed and validated over many years at Mason. They will run these simulations in parallel on their in-house supercomputing cluster.

Cebral received $115,000 from the U.S. Department of Health and Human Services for this work. Funding began in September 2020 and will end in August 2021.

MIT's machine learning discovers potential new TB drugs

A computational method for screening drug compounds can help predict which ones will work best against tuberculosis or other diseases

Machine learning is a computational tool used by many biologists to analyze huge amounts of data, helping them to identify potential new drugs. MIT researchers have now incorporated a new feature into these types of machine-learning algorithms, improving their prediction-making ability.

Using this new approach, which allows supercomputer models to account for uncertainty in the data they're analyzing, the MIT team identified several promising compounds that target a protein required by the bacteria that cause tuberculosis.

This method, which has previously been used by computer scientists but has not taken off in biology, could also prove useful in protein design and many other fields of biology, says Bonnie Berger, the Simons Professor of Mathematics and head of the Computation and Biology group in MIT's Computer Science and Artificial Intelligence Laboratory (CSAIL).

"This technique is part of a known subfield of machine learning, but people have not brought it to biology," Berger says. "This is a paradigm shift, and is absolutely how biological exploration should be done." {module INSIDE STORY}

Berger and Bryan Bryson, an assistant professor of biological engineering at MIT and a member of the Ragon Institute of MGH, MIT, and Harvard, are the senior authors of the study, which appears today in Cell Systems. MIT graduate student Brian Hie is the paper's lead author.

Better predictions

Machine learning is a type of computer modeling in which an algorithm learns to make predictions based on data that it has already seen. In recent years, biologists have begun using machine learning to scour huge databases of potential drug compounds to find molecules that interact with particular targets.

One limitation of this method is that while the algorithms perform well when the data they're analyzing is similar to the data they were trained on, they're not very good at evaluating molecules that are very different from the ones they have already seen.

To overcome that, the researchers used a technique called the Gaussian process to assign uncertainty values to the data that the algorithms are trained on. That way, when the models are analyzing the training data, they also take into account how reliable those predictions are.

For example, if the data going into the model predict how strongly a particular molecule binds to a target protein, as well as the uncertainty of those predictions, the model can use that information to make predictions for protein-target interactions that it hasn't seen before. The model also estimates the certainty of its own predictions. When analyzing new data, the model's predictions may have lower certainty for molecules that are very different from the training data. Researchers can use that information to help them decide which molecules to test experimentally.

Another advantage of this approach is that the algorithm requires only a small amount of training data. In this study, the MIT team trained the model with a dataset of 72 small molecules and their interactions with more than 400 proteins called protein kinases. They were then able to use this algorithm to analyze nearly 11,000 small molecules, which they took from the ZINC database, a publicly available repository that contains millions of chemical compounds. Many of these molecules were very different from those in the training data.

Using this approach, the researchers were able to identify molecules with very strong predicted binding affinities for the protein kinases they put into the model. These included three human kinases, as well as one kinase found in Mycobacterium tuberculosis. That kinase, PknB, is critical for the bacteria to survive but is not targeted by any frontline TB antibiotics.

The researchers then experimentally tested some of their top hits to see how well they actually bind to their targets and found that the model's predictions were very accurate. Among the molecules that the model assigned the highest certainty, about 90 percent proved to be true hits -- much higher than the 30 to 40 percent hit rate of existing machine learning models used for drug screens.

The researchers also used the same training data to train a traditional machine-learning algorithm, which does not incorporate uncertainty, and then had it analyze the same 11,000 molecule library. "Without uncertainty, the model just gets horribly confused and it proposes very weird chemical structures as interacting with the kinases," Hie says.

The researchers then took some of their most promising PknB inhibitors and tested them against Mycobacterium tuberculosis grown in bacterial culture media, and found that they inhibited bacterial growth. The inhibitors also worked in human immune cells infected with the bacterium.

A good starting point

Another important element of this approach is that once the researchers get additional experimental data, they can add it to the model and retrain it, further improving the predictions. Even a small amount of data can help the model get better, the researchers say.

"You don't really need very large data sets on each iteration," Hie says. "You can just retrain the model with maybe 10 new examples, which is something that a biologist can easily generate."

This study is the first in many years to propose new molecules that can target PknB, and should give drug developers a good starting point to try to develop drugs that target the kinase, Bryson says. "We've now provided them with some new leads beyond what has been already published," he says.

The researchers also showed that they could use this same type of machine learning to boost the fluorescent output of a green fluorescent protein, which is commonly used to label molecules inside living cells. It could also be applied to many other types of biological studies, says Berger, who is now using it to analyze mutations that drive tumor development.

ICDS's Roar supercomputer keeps an eye on volcano movements

RADAR satellites can collect massive amounts of remote sensing data that can detect ground movements, surface deformations, at volcanoes in near real-time. These ground movements could signal impending volcanic activity and unrest; however, clouds and other atmospheric and instrumental disturbances can introduce significant errors in those ground movement measurements.

Now, Penn State researchers have used artificial intelligence (AI) to clear up that noise, drastically facilitating and improving near real-time observation of volcanic movements and the detection of volcanic activity and unrest.

"The shape of volcanoes is constantly changing and much of that change is due to underground magma movements in the magma plumbing system made of magma reservoirs and conduits," said Christelle Wauthier, associate professor of geosciences and Institute for Data and Computational Sciences (ICDS) faculty fellow. "Much of this movement is subtle and cannot be picked up by the naked eye."

Geoscientists have used several methods to measure the ground changes around volcanoes and other areas of seismic activity, but all have limitations, said Jian Sun, lead author of the paper and a postdoctoral scholar in geosciences, funded by Dean's Postdoc-Facilitated Innovation through Collaboration Award from the College of Earth and Mineral Sciences.

He added that, for example, scientists can use ground stations, such as GPS or tiltmeters, to monitor possible ground movement due to volcanic activity. However, there are a few problems with these ground-based methods. First, the instruments can be expensive and need to be installed and maintained on site.

"So, it's hard to put a lot of ground-based stations in a specific area in the first place, but, let's say there actually is a volcanic explosion or an earthquake, that would probably damage a lot of these very expensive instruments," said Sun. "Second, those instruments will only give you ground movement measurements at specific locations where they are installed, therefore those measurements will have a very limited spatial coverage."

On the other hand, satellites and other forms of remote sensing can gather a lot of important data about volcanic activity for geoscientists. These devices are also, for the most part, out of harm's way from an eruption and the satellite images offer very extended spatial coverage of the ground movement. However, even this method has its drawbacks, according to Sun.

"We can monitor the movement of the ground caused by earthquakes or volcanoes using RADAR remote sensors, but while we have access to a lot of remote sensing data, the RADAR waves must go through the atmosphere to get recorded at the sensor," he said. "And the propagation path will likely be affected by that atmosphere, especially if the climate is tropical with a lot of water vapor and clouds variations in time and space."

According to the researchers, who report their findings in a recent issue of the Journal of Geophysical Research, a deep learning method they developed acts much like a jigsaw puzzle master. By taking pieces of data that are clear, the system can begin to fill in the holes of "noisy" data, holes created by the interference of weather, and other instrumental noises. It can then build a reasonably accurate picture of the land and its movements.

Using this deep learning method, scientists could gain valuable insights into the movement of the ground, particularly in areas with active volcanoes or earthquake zones and faults, said Sun. The program may be able to spot potential warning signs, such as sudden land shifts that might be a portent of an oncoming volcanic eruption, or earthquake.

"It's really important for areas close to active volcanoes, or near where there have been earthquakes, to have as early warning as possible that something might happen," said Sun.

Deep learning, as its name suggests, uses training data to teach the system to recognize features that the programmers want to study. In this case, the researchers trained the system with synthetic data that was similar to satellite surface deformation data. The data included signals of volcanic deformation, both spatially and topographically correlated atmospheric features, and errors in the estimation of satellite orbits.

Future research will focus on refining and expanding our deep learning algorithm, according to Wauthier.

"We wish to be able to identify earthquake and fault movements as well as magmatic sources and include several underground sources generating surface deformation," she said. "We will apply this new groundbreaking method to other active volcanoes thanks to support from NASA."