In this special guest feature, Shahbaz Ali of Tarmin takes a look at the commoditization of data and how aggregated data has evolved to the extent of being so crucial and proprietary that its value is unparalleled and irreplaceable to the organization.
In this special guest feature, Jesse Anderson from Cloudera writes about how many new companies, like the ones we see popping up in the Hadoop ecosystem, too quickly move from crawling to running, a process that sometimes leads to failure.
I recently ran across a thought-provoking post on the USC Anneberg Innovation Lab blog – “Why Do We Need Data Science when We’ve Had Statistics for Centuries.” With all the debate of late surrounding the relatively new “data science” term, I’ve been thinking a lot about this question, so I thought I’d analyze this notion […]
An eye-catching piece appearing in today’s edition of The Independent featured the thoughts of luminaries from the scientific world – renowned physicist Stephen Hawking, U.C. Berkeley computer-science professor Stuart Russell, and MIT physics professors Max Tegmark and Frank Wilczek – about the potential perils of artificial intelligence.
Unsupervised machine learning techniques have proven useful in identifying fake research papers submitted to the arXiv preprint server. Approximately 500 preprints are receiving daily by the automated repository arXiv, but are not pre-screened by humans. As a result, many nonsense papers generated by software such as SCIgen and Mathgen have been found in the most popular repository used by scientists to share research results.