Researchers at the Allen Institute have used machine learning to train computers to see parts of the cell the human eye cannot easily distinguish. Using 3D images of fluorescently labeled cells, the team taught computers to find structures inside living cells without fluorescent labels, using only black and white images generated by an inexpensive technique known as brightfield microscopy.
The study (“Label-free prediction of three-dimensional fluorescence images from transmitted-light microscopy”) is published in Nature Methods.
Fluorescence microscopy, which uses glowing molecular labels to pinpoint specific parts of cells, is very precise but only allows scientists to see a few structures in the cell at a time. Human cells have upwards of 20,000 different proteins that, if viewed together, could reveal important information about both healthy and diseased cells.
“Understanding cells as integrated systems is central to modern biology. Although fluorescence microscopy can resolve subcellular structure in living cells, it is expensive, slow, and can damage cells. We present a label-free method for predicting three-dimensional fluorescence directly from transmitted-light images and demonstrate that it can be used to generate multi-structure, integrated images. The method can also predict immunofluorescence (IF) from electron micrograph (EM) inputs, extending the potential applications,” write the investigators.
“This technology lets us view a larger set of those structures than was possible before,” says Greg Johnson, Ph.D., scientist at the Allen Institute for Cell Science, a division of the Allen Institute, and senior author on the study. “This means that we can explore the organization of the cell in ways that nobody has been able to do, especially in live cells.”
The prediction tool could also help scientists understand what goes wrong in cells during disease, says Rick Horwitz, Ph.D., executive director of the Allen Institute for Cell Science. Cancer researchers could potentially apply the technique to archived tumor biopsy samples to better understand how cellular structures change as cancers progress or respond to treatment. The algorithm could also aid regeneration medicine by uncovering how cells change in real time as scientists attempt to grow organs or other new body structures in the lab.
“This technique has huge potential ramifications for these and related fields,” Dr. Horwitz says. “You can watch processes live as they are taking place— it’s almost like magic. This method allows us, in the most noninvasive way that we have so far, to obtain information about human cells that we were previously unable to get.”
The combination of the freely available prediction toolset and brightfield microscopy could lower research costs if used in place of fluorescence microscopy, which requires expensive equipment and trained operators. Fluorescent tags are also subject to fading, and the light itself can damage living cells, limiting the technique’s utility to study live cells and their dynamics. The machine learning approach would allow scientists to track precise changes in cells over long periods of time, potentially shedding light on events such as early development or disease progression.
To the human eye, cells viewed in a brightfield microscope are sacs rendered in shades of gray. A trained scientist can find the edges of a cell and the nucleus, the cell’s DNA-storage compartment, but not much else. The research team used an existing machine learning technique, known as a convolutional neural network, to train computers to recognize finer details in these images, such as the mitochondria, cells’ powerhouses. They tested 12 different cellular structures and the model generated predicted images that matched the fluorescently labeled images for most of those structures, the researchers said.
It also turned out what the algorithm was able to capture surprised even the modeling scientists.
“Going in, we had this idea that if our own eyes aren’t able to see a certain structure, then the machine wouldn’t be able to learn it,” says Molly Maleckar, Ph.D., director of modeling at the Allen Institute for Cell Science and an author on the study. “Machines can see things we can’t. They can learn things we can’t. And they can do it much faster.”
The technique can also predict precise structural information from images taken with an electron microscope. The computational approach here is the same, says Forrest Collman, Ph.D., an assistant investigator at the Allen Institute for Brain Science, but the applications are different. Dr. Collman is part of a team working to map connections between neurons in the mouse brain. They are using the method to line up images of the neurons taken with different types of microscopes, normally a challenging problem for a computer and a laborious task for a human.
“Our progress in tackling this problem was accelerated by having our colleagues from the Allen Institute for Cell Science working with us on the solution,” Dr. Collman says.
Roger Brent, Ph.D., a member of the basic sciences division at Fred Hutchinson Cancer Research Center, is using the new approach as part of a research effort he is leading to improve the “seeing power” of microscopes for biologists studying yeast and mammalian cells.
“Replacing fluorescence microscopes with less light intensive microscopes would enable researchers to accelerate their work, make better measurements of cell and tissue function, and save some money in the process,” Dr. Brent says. “By making these networks available, the Allen Institute is helping to democratize biological and medical research.”