Big data - a 21st century science Maginot Line? No-boundary thinking: shifting from the big data paradigm
- Xiuzhen Huang1Email author,
- Steven F Jennings2,
- Barry Bruce3,
- Alison Buchan4,
- Liming Cai5,
- Pengyin Chen6,
- Carole L Cramer7,
- Weihua Guan8,
- Uwe KK Hilgert9,
- Hongmei Jiang10,
- Zenglu Li11,
- Gail McClure12,
- Donald F McMullen13,
- Bindu Nanduri14,
- Andy Perkins15,
- Bhanu Rekepalli16,
- Saeed Salem17,
- Jennifer Specker18,
- Karl Walker19,
- Donald Wunsch20,
- Donghai Xiong21,
- Shuzhong Zhang22,
- Yu Zhang23,
- Zhongming Zhao24 and
- Jason H Moore25Email author
© Huang et al.; licensee BioMed Central. 2015
Received: 17 December 2014
Accepted: 5 January 2015
Published: 6 February 2015
Whether your interests lie in scientific arenas, the corporate world, or in government, you have certainly heard the praises of big data: Big data will give you new insights, allow you to become more efficient, and/or will solve your problems. While big data has had some outstanding successes, many are now beginning to see that it is not the Silver Bullet that it has been touted to be. Here our main concern is the overall impact of big data; the current manifestation of big data is constructing a Maginot Line in science in the 21st century. Big data is not “lots of data” as a phenomena anymore; The big data paradigm is putting the spirit of the Maginot Line into lots of data. Big data overall is disconnecting researchers and science challenges. We propose No-Boundary Thinking (NBT), applying no-boundary thinking in problem defining to address science challenges.
KeywordsBig data Maginot Line No-Boundary thinking
The myth of big data
Big data has been largely promoted as a paradigm , bringing new challenges and opportunities. There are many national and international initiatives and funding programs [2,3] which focus on big data. The NIH definition of bioinformatics is essentially based on data: “Research, development, or application of computational tools and approaches for expanding the use of biological, medical, behavioral or health data, including those to acquire, store, organize, archive, analyze, or visualize such data ”. On the back cover of The Fourth Paradigm: Data-Intensive Scientific Discovery , Microsoft Corporation’s founder, Bill Gates, states “The impact of Jim Gray’s thinking is continuing to get people to think in a new way about how data and software are redefining what it means to do science.” There are many research projects and publications focused on big data; there are many big data-centered conferences and workshops; there are many big data hardware and software companies ; and there are many big data, high-throughput technologies, such as sequencing and imaging technologies. Big data seems to be getting more and more attention.
The Maginot Line
The French, based on the experiences of a previous generation during World War I, built a line of rigid fortifications along their border with Germany just prior to the start of World War II. While hailed as a work of genius at the time, by the time it was built, offensive military tactics turned it obsolete as the German’s Blitzkrieg simply bypassed those fortifications. By giving the French a false sense of security, the Maginot Line had the effect of draining resources from more flexible defensive strategies. The return on investment turned out to be poor. Is the big data paradigm turning out to be a science Maginot Line? Only time will tell, but we are already becoming aware of the limitations of this strategy.
The Cancer Genome Atlas (TCGA)
TCGA, with the goal to cover more than 20 different types of human cancers (>11,000 cases), is collecting data from different high-throughput platforms (including gene expression profiling, copy-number variation profiling, SNP genotyping, genome-wide DNA methylation profiling, microRNA profiling, and exon sequencing) and then releasing data usually after their analysis and publications. For the pilot project and phase II of TCGA, about US$200-million has been invested in this effort to gather samples, generate data, and analyze the data. TCGA publications, almost all in top-tier science journals and almost all with the similar titles as “Comprehensive Molecular Characterization of X Cancer” or “Comprehensive Molecular Profiling of Y Cancers,” for the most part present “stories” of their data generation and data analysis with some “plausible” results. If TCGA, with a comprehensive team of scientists and technology experts, could not dig the “gold” out of the collected large amount of data, how could other researchers be expected to do so? Furthermore, while collected data is static, the human genome is dynamic. So, should we continue collecting more and more data with the hope of digging out the “gold” information to save patients? Or, should we think about redirecting our efforts to specific, science-driven approaches, dynamic and systematic, to save dying patients from whom we collect the data? What is needed are not more reports, more lists of publications, more software packages, and more data. Efforts like TCGA are reaching the “bottleneck;” it is hard to make significant breakthroughs in scientific challenges by focusing on big data. Since interdisciplinary research does not work well, how about post-interdisciplinary approaches such as transdisciplinary approaches [communication with a senior scientist]? Since many current methods and approaches are generic, how about looking into more granular layers and finely-detailed approaches?
 NY Times: [Eight (No, Nine!) Problems With Big Data]: “Big data is here to stay, as it should be. But let’s be realistic: It’s an important resource for anyone analyzing data, not a silver bullet”.
 FT article: [Big data: are we making a big mistake?]: “Big datahas arrived, but big insights have not. The challenge now is to solve new problems and gain new answers – without making the same old statistical mistakes on a grander scale than ever”.
 RD article: [Why Big Data Isn't the Big Problem for Genomic Medicine]: “Of course, as this technology is adopted more broadly it will deliver new challenges in data management and analytics. But it’s nothing this industry can’t handle. The true barrier to clinical adoption of genomic medicine isn’t data volume or scale, but how to empower physicians from a logistical and clinical genomics knowledge standpoint, while proving the fundamental efficacy of genomics medicine in terms of improved patient diagnosis, treatment regimens, outcomes and improved patient management”.
Our main concern is not the ineffectiveness of big data for specific scientific problems. Also, our main concern is not for the numerous projects where big data seems to introduce significant false-positive results and potentially misleading discoveries (e.g., Cancer and chemotherapy are associated with a reduced Alzheimer’s risk ). Of course, specific projects may really need to collect big data to achieve the goals and to enable discoveries; our main concern here is not evaluating the need for big data in individual projects.
The Maginot Line gave France a false sense of security (since it seems strong and big); Is big data giving us a false sense of security, by assuming we could answer science challenges by looking at big data? The Maginot Line gave France a wrong impression of challenge (see how the German army attacked it); Is big data a real challenge? Big data may not be the challenge. It is the time we should re-focus on the science challenge, which is the real challenge.
We are proposing No-Boundary Thinking (NBT) to address real scientific challenges and to help science advance. Last year, we introduced the NBT concept . Rather than looking for big data or software tools to provide a connection among researchers of related disciplines, with NBT, the connection will come about by defining scientific problems to address science challenges. There are many problems based on big data approaches: not only is it just ineffective, but also it is disconnecting researchers from understanding the real science challenges. Currently the core of NBT is applying no-boundary thinking in problem defining.
NBT is not just adjusting the starting point from problem solving to problem defining, either. And it is not just starting earlier with interdisciplinary research. NBT is integrating life sciences and the computational and mathematical sciences closely and inseparably through no-boundary thinking. All researchers who bring similar and complementary interests and skills need to be integrated into problem defining as well as solving. NBT is also different from “multidisciplinary” or “transdisciplinary”; it is conceptualized without disciplinary limitations or boundaries (i.e., “discipline-free”). An article that explains these concepts and provides a detailed description of NBT uniqueness is in preparation.
Several decades ago with the boost of computers and software, there might have been a point to advocate for data and software for empowering science or to promote big data and software tools to connect researchers of different disciplines. However, today in the 21st century, the overall impact of the focus on big data and software is misleading and confusing to researchers and students, making their strategies rigid, which later on will have even broader negative impacts to science in science history.
Supported by NSF EPSCoR Grant Number #1239812 and NSF EAGER Grant Number #1452211. This work was also partially supported by the National Institute of Health grants from the National Center for Research Resources (P20RR016460) and the National Institute of General Medical Sciences (P20GM103429).
- The fourth paradigm: data-intensive scientific discovery, Tony Hey, Stewart Tansley and Kristin Tolle, Publisher: Microsoft Research, 1st ed. 2009.
- NIH BD2K [http://bd2k.nih.gov/], Big-Data to Knowledge (BD2K).
- NSF BIGDATA. [http://www.nsf.gov/cise/news/bigdata.jsp], the program of Critical Techniques and Technologies for Advancing Big Data Science & Engineering (BIGDATA).
- NIH Bioinformatics definition. [http://www.bisti.nih.gov/docs/CompuBioDef.pdf]
- E.g., ORACLE, SAS. [http://www.oracle.com/us/technologies/big-data/index.html]; [http://www.sas.com/en_us/insights/big-data/what-is-big-data.html]
- TCGA publications. [http://cancergenome.nih.gov/publications/TCGANetworkPublications]
- Eight (No, Nine!) problems with big data. Gary Marcus and Ernest Davis, New York Times, April 6, 2014.
- Big data: are we making a big mistake? Financial Times (FT) magazine, March 28, 2014.
- Why big data isn't the big problem for genomic medicine? Research & Development, February 14, 2014.
- Cancer and chemotherapy are associated with a reduced Alzheimer’s risk, Colby Stong, Editor. Neurol Rev. 2013; 21(11):1,22.
- Huang X, Bruce B, Buchan A, Congdon CB, Cramer CL, Jennings SF, et al. No-boundary thinking in bioinformatics research. BioData Min. 2013;6:19.View ArticlePubMedPubMed CentralGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly credited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.