The whisper of schizophrenia: Machine learning finds 'sound' words predict psychosis

A machine-learning method discovered a hidden clue in people's language predictive of the later emergence of psychosis

A machine-learning method discovered a hidden clue in people's language predictive of the later emergence of psychosis -- the frequent use of words associated with sound. A paper published by the journal npj Schizophrenia published the findings by scientists at Emory University and Harvard University.

The researchers also developed a new machine-learning method to more precisely quantify the semantic richness of people's conversational language, a known indicator for psychosis.

Their results show that automated analysis of the two language variables -- more frequent use of words associated with sound and speaking with low semantic density, or vagueness -- can predict whether an at-risk person will later develop psychosis with 93 percent accuracy. {module In-article}

Even trained clinicians had not noticed how people at risk for psychosis use more words associated with sound than the average, although abnormal auditory perception is a pre-clinical symptom.

"Trying to hear these subtleties in conversations with people is like trying to see microscopic germs with your eyes," says Neguine Rezaii, first author of the paper. "The automated technique we've developed is a really sensitive tool to detect these hidden patterns. It's like a microscope for warning signs of psychosis."

Rezaii began work on the paper while she was a resident at Emory School of Medicine's Department of Psychiatry and Behavioral Sciences. She is now a fellow at Harvard Medical School's Department of Neurology.

"It was previously known that subtle features of future psychosis are present in people's language, but we've used machine learning to actually uncover hidden details about those features," says senior author Phillip Wolff, a professor of psychology at Emory. Wolff's lab focuses on language semantics and machine learning to predict decision-making and mental health.

"Our finding is novel and adds to the evidence showing the potential for using machine learning to identify linguistic abnormalities associated with mental illness," says co-author Elaine Walker, an Emory professor of psychology and neuroscience who researches how schizophrenia and other psychotic disorders develop.

The onset of schizophrenia and other psychotic disorders typically occurs in the early 20s, with warning signs -- known as prodromal syndrome -- beginning around age 17. About 25 to 30 percent of youth who meet criteria for a prodromal syndrome will develop schizophrenia or another psychotic disorder.

Using structured interviews and cognitive tests, trained clinicians can predict psychosis with about 80 percent accuracy in those with a prodromal syndrome. Machine-learning research is among the many ongoing efforts to streamline diagnostic methods, identify new variables, and improve the accuracy of predictions.

Currently, there is no cure for psychosis. {module In-article}

"If we can identify individuals who are at risk earlier and use preventive interventions, we might be able to reverse the deficits," Walker says. "There are good data showing that treatments like cognitive-behavioral therapy can delay onset, and perhaps even reduce the occurrence of psychosis."

For the current paper, the researchers first used machine learning to establish "norms" for conversational language. They fed a computer software program the online conversations of 30,000 users of Reddit, a social media platform where people have informal discussions about a range of topics. The software program, known as Word2Vec, uses an algorithm to change individual words to vectors, assigning each one a location in a semantic space based on its meaning. Those with similar meanings are positioned closer together than those with far different meanings.

The Wolff lab also developed a computer program to perform what the researchers dubbed "vector unpacking," or analysis of the semantic density of word usage. Previous work has measured semantic coherence between sentences. Vector unpacking allowed the researchers to quantify how much information was packed into each sentence.

After generating a baseline of "normal" data, the researchers applied the same techniques to diagnostic interviews of 40 participants that had been conducted by trained clinicians, as part of the multi-site North American Prodrome Longitudinal Study (NAPLS), funded by the National Institutes of Health. NAPLS is focused on young people at clinical high risk for psychosis. Walker is the principal investigator for NAPLS at Emory, one of nine universities involved in the 14-year project. {module In-article}

The automated analyses of the participant samples were then compared to the normal baseline sample and the longitudinal data on whether the participants converted to psychosis.

The results showed that higher than normal usage of words related to sound, combined with a higher rate of using words with similar meaning, meant that psychosis was likely on the horizon.

Strengths of the study include the simplicity of using just two variables -- both of which have a strong theoretical foundation -- the replication of the results in a holdout dataset, and the high accuracy of its predictions, at above 90 percent.

"In the clinical realm, we often lack precision," Rezaii says. "We need more quantified, objective ways to measure subtle variables, such as those hidden within language usage."

Rezaii and Wolff are now gathering larger data sets and testing the application of their methods on a variety of neuropsychiatric diseases, including dementia.

"This research is interesting not just for its potential to reveal more about mental illness, but for understanding how the mind works -- how it puts ideas together," Wolff says. "Machine learning technology is advancing so rapidly that it's giving us tools to data mine the human mind."

Portuguese researcher develops new model that more accurately predicts choices in classic decision-making task

Exploratory behaviors in the Iowa Gambling Task seem to decline with aging

A new mathematical model that predicts which choices people will make in the Iowa Gambling Task, a task used for the past 25 years to study decision-making, outperforms previously developed models. Romain Ligneul of the Champalimaud Center for the Unknown in Portugal presents this research in PLOS Computational Biology.

The Iowa Gambling Task presents a subject with four virtual card decks, each containing a different mix of cards that can win or lose fake money. Without being told which decks are more valuable, the subject then picks cards from the decks as they please. Most healthy people gradually learn which decks are more valuable and choose to pick cards only from those decks.

Earlier studies have used Iowa Gambling Task data to build mathematical models that can predict people's card-picking choices. However, building such models is computationally challenging, and previously developed models do not account for the exploratory strategies people use in the task. CAPTION Which door will you choose? New model helps predict what choices people make in the Iowa Gambling Task by focusing on the 'exploratory strategies' they use.  CREDIT dil/unsplash.{module In-article}

In reviewing previously collected data from 500 subjects, Ligneul found that healthy people tend to cycle through the four decks and pick one card from each, especially at the beginning of the task. He then incorporated this behavior, termed sequential exploration, into a new mathematical model that also accounts for the well-known reward-maximizing behaviors people exhibit in the task.

Ligneul found that his new model outperforms earlier models in predicting people's card-picking choices. He also found that sequential exploration behaviors seem to decline as subjects get older, perhaps because of neurological changes typically associated with aging.

"This study provides a mathematical method to disentangle our drive to explore the environment and our drive to exploit it," Ligneul says. "It appears that the balance of these two drives evolves with aging."

The new model and findings could help refine insights gleaned from the Iowa Gambling Task. It could also improve understanding of learning and decision-making disruptions that are associated with aging and various neuropsychiatric conditions, such as addiction, impulsive disorders, brain injury, and more.

Artificial intelligence beats physicians in the diagnosis of skin lesions

When it comes to the diagnosis of pigmented skin lesions, artificial intelligence is superior to humans. In a study conducted under the supervision of the MedUni Vienna human experts "competed" against computer algorithms. The algorithms achieved clearly better results, yet their current abilities cannot replace humans. 

The International Skin Imaging Collaboration (ISIC) and the MedUni Vienna organized an international challenge to compare the diagnostic skills of 511 physicians with 139 computer algorithms (from 77 different machine learnings labs). A database of more than 10.000 images, which was established by the team around Harald Kittler at the Department of Dermatology of MedUni Vienna in cooperation with the University of Queensland (Australia), was used as a training set for the machines. This database includes benign (moles, sun spots, senile warts, angiomas, and dermatofibromas) and malignant pigmented lesions (melanomas, basal cell carcinoma, and pigmented squamous cell carcinoma). {module In-article} 

Each participant had to diagnose 30 randomly selected images out of a test-set of 1511 images. The result was unequivocal. While the best humans diagnosed 18.8 out of 30 cases correctly, the best machines achieved 25.4 correct diagnoses. This did not surprise first-author Philipp Tschandl from the MedUni Vienna: "Two-thirds of all participating machines were better than humans; this result had been evident in similar trials during the past years."

Not a substitute for human beings

Although the algorithms were clearly superior in this experiment, this does not mean that the machines will replace humans in the diagnosis of skin cancer. Philipp Tschandl: "The computer only analyzes an optical snapshot and is really good at it. In real life, however, the diagnosis is a complex task. Physicians usually examine the entire patient and not just single lesions. When humans make a diagnosis they also take additional information into account, such as the duration of the disease, whether the patient is at high or low risk, and the age of the patient, which was not provided in this study.

Despite the impressive performance of artificial intelligence, there is still room for improvement. The machines were significantly less accurate in the diagnosis of lesions that came from centers that did not provide training images. With regard to human performance, the experience was important. The most experienced participants with at least ten years of experience in the diagnosis of pigmented skin lesions performed best. The results were published in the journal “The Lancet Oncology."