Steve Conway, IDC VP HPC explains that, to date, most data-intensive HPC jobs in the government, academic and industrial sectors have involved the modeling and simulation of complex physical and quasi-physical systems. However, he notes that from the start of the supercomputer era in the 1960s — and even earlier — an important subset of HPC jobs has involved analytics, attempts to uncover useful information and patterns in the data itself.
This month’s review is a bit off of the usual track, e.g. statistical, mathematical and genomics software. However, it does include much pertinent information for chemists, chemical engineers and biologists. SciFinder is a search engine for chemistry and biology references for just about anything that can be accurately described in the search feature.
Welcome to Informatics Snapshot — a feature that highlights the standout properties of the current crop of laboratory informatics systems. While not intended to be a full formal review of the featured product or to indicate whether the product is considered “good” or “bad,” its purpose is to present some of the “diamonds and rust,” as the Joan Baez song goes. In this article, we’ll take a brief look at the LabX system
Big Data tools such as Grok and IBM Watson are enabling large organizations to behave more like agile startups. Of the transformative technology developments that have ushered in the current frenzy of activity along the information superhighway, the 1994 invention of the “Wiki” by Ward Cunningham is among the most disruptive.
Encryption and nuclear weapons are two easily recognized examples where a combinatorial explosion is a sought after characteristic. In the software development world, combinatorial explosions are bad. In particular, it is far too easy to become lost in the minutia of writing code that can run efficiently on NVIDIA GPUs, AMD GPUs, x86, ARM and Intel Xeon Phi while also addressing the numerous compiler and user interface vagaries
Data Integrity in a Nutshell: Industry must take bold steps to assure the data used for drug quality decisions is trustworthyJanuary 7, 2014 12:31 pm | by Mark E. Newton | Comments
Regulatory inspectors have started digging much deeper into data, no longer accepting batch release data and supportive testing at face value. Even worse, this effort is justified: they have cited a number of firms for violations of data integrity, a most fundamental bond of trust between manufacturers and the regulators that inspect them. Industry must take bold steps to assure the data used for drug quality decisions is trustworthy...
Data integrity is a current hot topic with regulatory agencies, as seen with recent publications in this magazine, and audit trails are an important aspect of ensuring this in computerized systems. The purpose of this article is to compare and contrast the EU and FDA GMP regulatory requirements for computerized system audit trails.
One of the challenges in laboratory data management is the handling and exchange of experiment data. Many vendors provide excellent instruments, but most produce data in their own proprietary formats. This leads to major difficulties for data processing, collaboration, instrument integration and archiving. The ASTM AnIML standardization effort addresses these problems by providing a neutral XML-based format for exchanging scientific data.
Mobile technology is where the money is right now in computer technology. Current leadership class supercomputers are “wowing” the HPC world with petaflop/s performance through the combined use of several thousand GPUs or Intel Xeon Phi coprocessors, but in reality the sale of a few thousand of these devices is insignificant when compared against the 1.5 billon cellphone processors and 190 million tablet processors ...
Software Review: Unscrambler statistical software is geared to two of the most useful areas of industrial R&D, namely multivariate analysis and experimental design. The latest version (10.3) of this useful niche software has a number of additions and upgrades, including regression and classification methods, exploratory data analysis tools, predictive modeling, extensive pre-processing options, and descriptive statistics with tests.
One of the most magical aspects of mathematics is the ability for complex and intricate structures to emerge from the simplest of rules. Few mathematical objects are simpler to create — and few weave such a variety of intricate patterns — as Pascal’s marvellous triangle.
After unexpectedly missing the opportunity to exhibit their expertise at SC12, the Department of Energy (DOE) national laboratories will return to the conference exhibition at the SC13 international conference for high performance computing, networking, storage and analysis, to be held November 17 to 22 at the Colorado Convention Center (CCC) in Denver.
Change is a given in the technology world as new products excite interest, generate sales and, ultimately, define profitability. No technology company is “too big to fail,” which means that the current market giants recognize they can easily become a name from the past, like SUN Microsystems and Digital Equipment Corporation, unless they aggressively innovate.
Everything leading up to the actual coding, figuring out how to make it work, is what Samak enjoys most. One of the problems she is working on with the Department of Energy’s Joint Genome Institute (JGI) is a data mining method to automatically identify errors in genome assembly, replacing the current approach of manually inspecting the assembly.
Taghrid Samak of Berkeley Lab’s Computational Research Division admits with a laugh that she wasn’t one of those kids who started programming on the home computer at age 10. And if she hadn’t followed her father’s advice, she might have ended up looking for political solutions to pressing problems, rather than working on computational approaches to scientific challenges.