Researchers have found a way for computers to recognize 21 distinct facial expressions - even expressions for complex or seemingly contradictory emotions such as “happily disgusted” or “sadly angry.”
Ryan Kennedy, University of Houston political science professor, and his co-researchers detail...
Ben Recht is looking for problems. He develops mathematical strategies to help researchers, from...
Next week, Scientific Computing will host a live panel discussion that looks at how a...
NASA’s Asteroid Data Hunter contest series will offer $35,000 in awards over the next six months to citizen scientists who develop improved algorithms that can be used to identify asteroids. This contest is being conducted in partnership with Planetary Resources of Bellevue, WA.
How can organizations embrace — instead of brace for v the rapidly intensifying collision of public and private clouds, HPC environments and Big Data? The current go-to solution for many organizations is to run these technology assets in siloed, specialized environments. This approach falls short, however, typically taxing one datacenter area while others remain underutilized, functioning as little more than expensive storage space.
At Cycle Computing we’re seeing several large trends as it relates to Big Data and Analytics. We started talking about this concept of Big Compute back in Oct. 2012. In many ways, it’s the collision of where HPC is meeting the challenges of Big Data. As our technical capabilities continue to expand in the ways we can collect and store data, the problem of how we access and use data is only growing.
Scalable Productivity and the Ever-Increasing Tie between Data Analytics, Data Management and ComputationMarch 7, 2014 3:52 pm | by Barry Bolding, Cray | Blogs | Comments
Cray continues to see an increasing trend in the HPC marketplace that we are calling “data-intensive” supercomputing. The dramatic growth in scientific, commercial and social data is resulting in an expanded customer base that is asking for much more complex analysis and simulation.
In 2013, the term big data continued to dominate as a source of technology challenges, experimentation and innovation. It’s no surprise then that many business and IT executives are suffering from big data exhaustion, causing Gartner to deem 2013 as the year the technology entered the “Trough of Disillusionment.”
From the start of the supercomputer era in the 1960s — and even earlier —an important subset of HPC jobs has involved analytics — attempts to uncover useful information and patterns in the data itself. Cryptography, one of the original scientific-technical computing applications, falls predominantly into this category.
Steve Conway, IDC VP HPC explains that, to date, most data-intensive HPC jobs in the government, academic and industrial sectors have involved the modeling and simulation of complex physical and quasi-physical systems. However, he notes that from the start of the supercomputer era in the 1960s — and even earlier — an important subset of HPC jobs has involved analytics, attempts to uncover useful information and patterns in the data itself.
Size alone does not define big data — it is best defined as a combination of volume, velocity, variety and value.Kevin Geraghty, head of Analytics 360i defined the goal of big data analytics well when he said: “We are trying to listen to what the customer is telling us through their behavior.” The goal of big data analytics is to make the best business decisions possible.
NASA scientists have begun deploying satellites and other advanced technology to help California water officials assess the state's record drought and better manage it, officials said February 25, 2104. The California Department of Water Resources has partnered with NASA to use the space agency's satellite data and other airborne technology to better measure the snowpack, groundwater levels and predict storms.
AT&T and IBM have announced a new global alliance agreement to develop solutions that help support the "Internet of Things." The companies will combine their analytic platforms, cloud and security technologies with privacy in mind to gain more insights on data collected from machines in a variety of industries.
The Intel Xeon processor E7 v2 family delivers capabilities to process and analyze large, diverse amounts of data to unlock information that was previously inaccessible. The processor family has triple the memory capacity of the previous generation processor family, allowing much faster and thorough data analysis.
HPCS is a multi-disciplinary conference, considered Canada's premier advanced computing forum. Each year, Canadian researchers, analysts, and IT professionals from academia and industry gather to exchange the ideas, tools, and new discoveries that are driving today's innovations in computational research.
IBM has launched a 10-year initiative to bring Watson and other cognitive systems to Africa in a bid to fuel development and spur business opportunities across the world's fastest growing continent. Dubbed "Project Lucy" after the earliest known human ancestor, IBM will invest US$100 million in the initiative
As computers enter ever more areas of our daily lives, the amount of data they produce has grown enormously. But for this “big data” to be useful it must first be analyzed, meaning it needs to be stored in such a way that it can be accessed quickly when required.
Texas A&M System Teams with IBM to Drive Computational Sciences Research through Big Data and AnalyticsJanuary 29, 2014 1:38 pm | by IBM | News | Comments
Texas A&M University System and IBM announced an agreement that is the beginning of a broad research collaboration supported by one of the largest computational sciences infrastructures dedicated to advances in agriculture, geosciences and engineering.
Inspired by nature, scientists from Berlin and Heidelberg use artifical nerve cells to classify different types of data. A bakery assistant who takes the bread from the shelf just to give it to his boss who then hands it over to the customer? Rather unlikely. Instead, both work at the same time to sell the baked goods.
Millions of American families have already embraced connected devices like smart meters to control their electricity bills. Now San Diego will soon become the first U.S. city to link its street lights to the Industrial Internet via an “intelligent” lighting system called LightGrid. The new system, which was developed by GE Lighting, will replace 3,000 city lamps with LED lights equipped with...
From April 1 – 4, the International Trade Fair for Laboratory Technology, Analysis and Biotechnology will be a center for key players in science and industry. This year’s analytica will revolve around three main themes — i.e. food analysis, plastics analysis and genetic and bioanalysis—whether in the exhibition, the Live Labs and the program of related events.
Are you a B.Sc. or M.Sc. student in computer science, mathematics, engineering or physics, interested in working on advanced IT projects at CERN for two full months during the summer? CERN openlab has R&D activities on multithreaded programming, data acquisition systems, large scale databases, data analytics, and cloud computing that are open to students who will have completed at least three years of full-time studies.
This month, the Texas Advanced Computer Center (TACC) at The University of Texas at Austin, US, along with technology partners HP and NVIDIA, will deploy Maverick — a powerful, high-performance visualization and data analytics resource for the open science and engineering community.
GE Intelligent Platforms announced that it has received an order from the High Performance Systems Branch (RITB) of the US Air Force Research Laboratory (AFRL) Information Directorate (RI) for a HPEC (High Performance Embedded Computing) system that will enable the development and deployment of advanced neuromorphic architectures and algorithms for adaptive learning, large-scale dynamic data analytics and reasoning.
IBM has unveiled three new Watson services delivered over the cloud. The first, Watson Discovery Advisor, is designed to accelerate and strengthen R&D projects in industries such as pharmaceutical, publishing and biotech. The second, Watson Analytics, delivers visualized Big Data insights, based on questions posed in natural language. The third, Watson Explorer, helps users across an enterprise uncover and share data-driven insights
University College Dublin has honored the co-founders of Logentries, a UCD 'big data' spin-out company by presenting them with the NovaUCD 2013 Innovation Award.
Mellanox Technologies, a supplier of high-performance, end-to-end interconnect solutions for data center servers and storage systems, announced that its dual port ConnectX-3 10/40GbE Network Interface Cards (NICs) are now fully compatible with qualified Dell PowerEdge servers and Dell networking solutions, providing improved performance at higher efficiency and lower latency with reduced application costs.
The Texas Advanced Computing Center (TACC) at The University of Texas at Austin has announced the addition of Rodeo, the center's new cloud computing offering that will provide a variety of cloud services to its comprehensive set of advanced computing systems. The system is available for researchers that need large-scale resources, but prefer a virtualized, on-demand environment.
- Page 1