MIT spinout Akselos has developed novel software, based on years of research at the Institute, which uses precalculated supercomputer data for structural components — like simulated “Legos” — to solve FEA models in seconds. Hundreds of engineers in the mining, power-generation, and oil and gas industries are now using Akselos software.
The Michael J. Fox Foundation for Parkinson's Research (MJFF) and Intel have announced a...
Scientists from IBM have unveiled the first neurosynaptic computer chip to achieve an...
As our lives and businesses become ever more intertwined with the Internet and networked...
The PrecisionMDx (PMDx) System is designed for molecular laboratories. It supports molecular diagnostic requirements of entrepreneurial molecular test development companies and the institutional services of molecular pathology departments in community, private and academic medical centers.
Scientists from AT&T, IBM and ACS announced a proof-of-concept technology that reduces set up times for cloud-to-cloud connectivity from days to seconds. This advance is a major step forward that could one day lead to sub-second provisioning time with IP and next-generation optical networking equipment and enables elastic bandwidth between clouds at high connection request rates using intelligent cloud data center orchestrators.
Enabling Innovation and Discovery through Data-Intensive High Performance Cloud and Big Data InfrastructureJuly 29, 2014 2:34 pm | by George Vacek, DataDirect Networks | Blogs | Comments
As the size and scale of life sciences datasets increases — think large-cohort longitudinal studies with multiple samples and multiple protocols — so does the challenge of storing, interpreting and analyzing this data. Researchers and data scientists are under increasing pressure to identify the most relevant and critical information within massive and messy data sets, so they can quickly make the next discovery.
In an age of “big data,” a single computer cannot always find the solution a user wants. Computational tasks must instead be distributed across a cluster of computers that analyze a massive data set together. It's how Facebook and Google mine your Web history to present you with targeted ads, and how Amazon and Netflix recommend your next favorite book or movie. But big data is about more than just marketing.
IBM is making high performance computing more accessible through the cloud for clients grappling with big data and other computationally intensive activities. A new option from SoftLayer will provide industry-standard InfiniBand networking technology to connect SoftLayer bare metal servers. This will enable very high data throughput speeds between systems, allowing companies to move workloads traditionally associated with HPC to the cloud.
A case study published in The International Journal of Business Process Integration and Management demonstrates that the adoption of integrated cloud-computing solutions can lead to significant cost savings for businesses, as well as large reductions in the size of an organization's carbon footprint.
The National Institute of Standards and Technology (NIST) has issued for public review and comment a draft report summarizing 65 challenges that cloud computing poses to forensics investigators who uncover, gather, examine and interpret digital evidence to help solve crimes.
IBM is announcing a new software defined storage-as-a-service on IBM SoftLayer, code named Elastic Storage on Cloud, that gives organizations access to a fully-supported, ready-to-run storage environment, which includes SoftLayer bare metal resources and high performance data management and allows organizations to move data between their on-premise infrastructure and the cloud.
Registration is now open for the 2014 ISC Cloud and ISC Big Data Conferences, which will be held this fall in Heidelberg, Germany. The fifth ISC Cloud Conference will take place in the Marriott Hotel from September 29 to 30, and the second ISC Big Data will be held from October 1 to 2 at the same venue.
Michael M. Resch, the Director of the Stuttgart High Performance Computing Center (HLRS) will be talking about “HPC and Simulation in the Cloud – How Academia and Industry Can Benefit.” His keynote is of special interest to cloud skeptics, given that prior to 2011, Resch himself was a vocal cloud pessimist. Three years later, he feels that this technology provides a practical option for many users.
IBM Announces $3B Research Initiative to Tackle Chip Grand Challenges for Cloud and Big Data SystemsJuly 9, 2014 4:58 pm | by IBM | News | Comments
IBM has announced it is investing $3 billion over the next five years in two broad research and early stage development programs to push the limits of chip technology needed to meet the emerging demands of cloud computing and Big Data systems. These investments are intended to push IBM's semiconductor innovations from today’s breakthroughs into the advanced technology leadership required for the future.
Moab HPC Suite-Enterprise Edition 8.0 (Moab 8.0) is designed to enhance Big Workflow by processing intensive simulations and big data analysis to accelerate insights. It delivers dynamic scheduling, provisioning and management of multi-step/multi-application services across HPC, cloud and big data environments. The software suite bolsters Big Workflow’s core services: unifying data center resources, optimizing the analysis process and guaranteeing services to the business.
An energy efficient supercomputer with warm water. How cool is that? Enlightenment has long been the ultimate pursuit of artists, philosophers, scientists, theologians and other sentient minds. Whether it is delivering the proof to support their theses, or to investigate a perplexing problem before them, they have poured a vast amount of energy into the situation. Energy has now become the problem.
HP has announced new innovations and sustainable enterprise infrastructure solutions designed to deliver the simplicity, efficiency and investment protection organizations need to bridge the datacenter technologies of today and tomorrow. Big data, mobility, security and cloud computing are forcing organizations to rethink their approach to technology, causing them to invest heavily in IT infrastructure.
The lack of a holistic data management environment to support virtualization has left project managers in a haze about how best to address the needs of the business. The sky is beginning to clear somewhat with recent introductions from companies such as Accelrys, Core Informatics and PerkinElmer. Those products, along with CDD, will be discussed to highlight capabilities and vendor approaches.
A complicated decision: To purchase infrastructure or run remotely in the cloud? Bandwidth and data security issues provide the easiest gating factors to evaluate, because an inability to access data kills any chance of using remote infrastructure, be it the public cloud or at a remote HPC center. If running remotely is an option, then the challenge lies in determining the return on investment (ROI) for the remote and local options ...
Atos, an international information technology services company, and Bull, a partner for enterprise data, together announced the intended public offer in cash by Atos for all the issued and outstanding shares in the capital of Bull. Atos offer is set at 4.90 euros per Bull's share in cash, representing a 22 percent premium over the Bull's closing price
Elastic Storage is capable of reducing storage costs up to 90 percent by automatically moving data onto the most economical storage device. The technology allows enterprises to exploit the exploding growth of data in a variety of forms generated by devices, sensors, business processes and social networks.
Penguin Computing, experts in high performance, enterprise and cloud computing solutions, has announced the immediate availability of MATLAB Distributed Computing Server on its HPC Cloud, POD. This solution combines POD’s ease-of-use and high performance computing capabilities in the cloud with MATLAB scale-up capability to solve more demanding and complex problems.
At EMC World 2014, EMC announced major new software-defined storage products and technologies designed to enable the blended benefits of a public and private cloud, delivering service providers and users in any industry and of any size, the efficiency, agility, security and control of a hybrid cloud.
Governor Deval Patrick announced a $3 million capital investment to launch the Massachusetts Open Cloud project, a university-industry collaboration designed to create a new public cloud computing infrastructure to spur big data innovation, on April 25, 2014. Governor Patrick also announced the release of the 2014 Mass Big Data Report, which confirms the continued growth and competitiveness of the Commonwealth’s big data industry.
This year’s ISC Cloud conference will be the fifth in the series and will continue the tradition of bringing experts and users from industry and academia to foster collaboration and innovation in the field of cloud computing. The conference will be held in Heidelberg at the Marriott Hotel from September 29 to 30, 2014, followed by ISC Big Data
Sharan Kalwani recently joined the HPC group at the Fermi National Accelerator Labs, Batavia, Illinois as a computing services architect. Before Fermi, he was the Subject Matter Expert/Project lead at the UberCloud project, working on helping to realize HPC in the cloud.
Muniyappa Manjunathaiah's research in computational science includes novel and emergent systems and architectures, parallel and distributed computing, cloud computing, mathematical modelling, scalable algorithms, middleware to support parallel and distributed applications.
- Page 1