Challenge to be first in a series to crowdsource Alzheimer’s data for new disease insights .
Can Big Data be used to foster serendipity? That’s the premise of an award-winning paper in the 2013 Semantic Web Challenge. Entitled “Fostering Serendipity through Big Linked Data” the paper was written by Muhammad Saleem, Maulik R. Kamdar, Aftab Iqbal, Shanmukha Sampath, Helena F. Deus and Axel-Cyrille Ngonga Ngomo. The amount of bio-medical data available […]
In this video, Steve Forbes hosts Sir Andrew Witty from GlaxoSmithKline in a discussion on how data and data sharing is having a profound impact on people’s health around the world. Traditionally, big pharma kept their data and results under lock and key. Now, with a global population impatient to find answers and cures, the industry is working together to affect change.
On Tuesday, April 2, President Obama announced a research initiative that has the ambitious goal of “revolutionizing our understanding of the human brain,” according to a White House press release. Know as BRAIN (Brain Research through Advancing Innovative Neurotechnologies), the initiative is being launched in FY 2014 with an initial budget of about $100 million, […]
In this special guest feature, Anchita Magan from [x]cube DATA writes that the element of quality has to be considered in quantifiable data. Significance of Big Data The entire cosmos has been turned into an aggregated ocean of Data – structured or unstructured, systematic or unsystematic, useful or useless. This zillion of roughly organized data […]
You can’t find better examples of the promises and pitfalls of Big Data than in the realm of the life sciences. According to an excellent article, “Unraveling the Complexities of Life Sciences Data,” published in the journal Big Data, with the combination of the completion of the human genome project and the availability of advanced […]
You have to wonder what Albert Einstein would be up to in this era of big data. Writing in Amdahl’s and
In this video from SC12, Steve Simms from Indiana University describes a recent upgrade to the Data Capacitor project, a high-speed, high-capacity storage facility for very large data sets. With 5 PB of storage, Data Capacitor II will support big data applications used in computational research. IU partnered with DataDirect Networks to develop Data Capacitor […]