Data-intensive Computing
            
			
          Data-Intensive Computing Research and Education
          Project partially funded by National Science Foundation Grant
          NSF-DUE-CCLI-0920335: PI B. Ramamurthy, $249K.
		  
		  
		  
An Excerpt from the funded grant (2009-2014)
		  
		  Data-intensive computing has been receiving much attention as a collective
          solution to address the data deluge that has been brought
          about by tremendous advances in distributed systems and
          Internet-based computing. An innovative programming models
          such as MapReduce and a peta-scale distributed file system to
          support it have revolutionized and fundamentally changed
          approaches to large scale data storage and processing.
		  These data-intensive computing
          approaches are expected to have profound impact on any
          application domain that deals with large scale data, from
          healthcare delivery to military intelligence. A new forum
          called big-data computing group
          has been formed by a consortium of industrial stakeholders and
          agencies including NSF and CRA (Computing Research Associates)
          to promote wider dissemination of the big-data solutions and
          transform mainstream applications. Given the omnipresent
          nature of large scale data and the tremendous impact they have
          on a wide variety of application domains, it is imperative to
          ready our workforce to face the challenges in this area. This project aims to improve
          the big-data preparedness of diverse STEM
          audience by defining a comprehensive framework for education
          and research in data-intensive computing.
		  
A new certificate program in data-intensive computing
          has been
                approved by SUNY and is offered by the University. 
                The details are in the 
                  catalog page here.