Over a million developers have joined DZone.
{{announcement.body}}
{{announcement.title}}

Data-Intensive Computing to Simulate the Brain

DZone's Guide to

Data-Intensive Computing to Simulate the Brain

Along with making inroads into neuroscience and treating brain disease, this research could power tomorrow’s technologies — from our computers to artificial intelligence.

· AI Zone
Free Resource

Find out how AI-Fueled APIs from Neura can make interesting products more exciting and engaging. 

Understanding how the human brain works will take more than brains. Along with the planet’s smartest scientific minds, it will take never-before-achieved computing capabilities.

The science and technology required to decode the human brain is a scientific final frontier — and Professor Dr. Dirk Pleiter is on the front lines. The theoretical physics professor and research group leader at the Jülich Supercomputing Centre (JSC) in Jülich, Germany, is part of the Human Brain Project (HBP), a 10-year-long European research initiative tasked with creating a working simulation of the brain.

“Understanding the human brain is one of the greatest challenges facing 21st-century science,” states the HBP’s report to the European Commission. “If we can rise to the challenge, we can gain fundamental insights into what it means to be human, develop new treatments for brain diseases, and build revolutionary new information and communications technologies.

This ambitious project requires a leap in scientific research and technological capabilities beyond what exists today. It demands to bring together high-performance computing and data analysis as tools in neuroscience — an effort that Pleiter describes as “pioneering.”

“The human brain is the most complex system we know of,” says Pleiter. “We’re developing a view of the brain where we can zoom in and out, see it from different perspectives and understand how brain structure and function are related.”

Computationally speaking, the task has diverse requirements. Modeling the human brain starts as a simulation challenge and quickly turns into a big data challenge. Neuroscience supercomputers will have to scale and handle data volumes surpassing what they do today. “We’ll need tens of petabytes of memory if researchers want to be able to scale their models to the size of the human brain,” he says.

Dr. Pleiter and the JSC team are hard at work on it, developing the data-intensive supercomputing infrastructure that will allow researchers to make advancements in neuroscience, brain-related medicine and even computing. Step one is pilot systems. Cray delivered its first one to JSC, and the team is using it to develop and test simulation and analysis techniques. Ultimately, scientists will need to be able to interactively visualize and control large-scale simulations.

The project’s future applications are vast. Along with making inroads into neuroscience and treating brain disease, scientists expect this research could power tomorrow’s technologies — from our computers to artificial intelligence.

Says Pleiter:

“This true combining of big data problems and HPC…will help us go in new directions and push science forward.”

To find out how AI-Fueled APIs can increase engagement and retention, download Six Ways to Boost Engagement for Your IoT Device or App with AI today.

Topics:
ai ,data science ,machine learning ,computing

Published at DZone with permission of Adnan Khaleel, DZone MVB. See the original article here.

Opinions expressed by DZone contributors are their own.

{{ parent.title || parent.header.title}}

{{ parent.tldr }}

{{ parent.urlSource.name }}