We are testing a new system for linking grants to scientists.
The funding information displayed below comes from the
NIH Research Portfolio Online Reporting Tools and the
NSF Award Database.
The grant data on this page is limited to grants awarded in the United States and is thus partial. It can nonetheless be used to understand how funding patterns influence mentorship networks and vice-versa, which has deep implications on how research is done.
You can help! If you notice any innacuracies, please
sign in and mark grants as correct or incorrect matches.
Sign in to see low-probability grants and correct any errors in linkage between grants and researchers.
High-probability grants
According to our matching algorithm, David Waltz is the likely recipient of the following grants.
Years |
Recipients |
Code |
Title / Keywords |
Matching score |
2005 — 2009 |
Waltz, David L |
U54Activity Code Description: To support any part of the full range of research and development from very basic to clinical; may involve ancillary supportive activities such as protracted patient care necessary to the primary research or R&D effort. The spectrum of activities comprises a multidisciplinary attack on a specific disease entity or biomedical problem area. These differ from program project in that they are usually developed in response to an announcement of the programmatic needs of an Institute or Division and subsequently receive continuous attention from its staff. Centers may also serve as regional or national resources for special research purposes, with funding component staff helping to identify appropriate priority needs. |
Core--Computational Sciences @ Columbia University Health Sciences |
0.958 |
2009 — 2013 |
Waltz, David Dutta, Haimonti Schevon, Catherine Emerson, Ronald |
N/AActivity Code Description: No activity code was retrieved: click on the grant title for more information |
Dc: Small: Eegmine: a Distributed Framework For Learning On Eeg Data Obtained From Epilepsy Patients
The Center for Computational Learning Systems (CCLS) is collaborating with the Computational Neurophysiology Laboratory (CNL) in the Department of Neurology, Columbia University Medical School (CUMC) to develop a distributed framework for data management and machine learning on intracranial EEG data obtained from patients suffering from epilepsy.
Drs. Schevon and Emerson have initiated a trial of a dense, two-dimensional microelectrode array which can record over long periods of time at a sampling rate of up to 30 kHz per channel. To date approximately 30 TB of data has been collected. The large volume of complex EEG data compels us to rethink how we will deal with this "data avalanche." The design of a data center for storage and analysis is particularly challenging since traditional methods of storing data on a single server do not allow machine learning algorithms to be computed within a reasonable time. Further, due to the conditions under which the data is collected, noise of multiple types and sources is pervasive; the data must be extensively cleaned and potential seizure precursors carefully labeled. The project is investigating mechanisms to develop a cluster architecture (using Apache Hadoop) for the EEGMine Data Center that incorporates reliable storage and backup; developing a library of machine learning algorithms (EEGMine- ML library) and addressing their scalability issues, potentially leveraging the MapReduce programming paradigm.
This research will have immediate impact for both epilepsy and computer science research. Because of the uniqueness and value of human-derived microelectrode EEG data, it would be beneficial for the seizure prediction community to enable data sharing and long-distance collaborations. The most practical means of sifting through terabytes of complex EEG data is to combine distributed storage on a cluster with local processing to prepare data and generate meta-data that can be used as inputs for machine learning algorithms thus enabling identification of physiologically significant patterns. From an education perspective, the project will benefit the EWarn Research Group which is part of CCLS and CUMC by training them in signal processing, machine learning and basics of EEG.
Website Address: http://www1.ccls.columbia.edu/~dutta/EEGMine
|
1 |
2009 — 2011 |
Waltz, David Vapnik, Vladimir (co-PI) [⬀] |
N/AActivity Code Description: No activity code was retrieved: click on the grant title for more information |
Ri: Small: An Advanced Learning Paradigm: Learning Using Hidden Information
Modern machine learning is limited in its ability to use diverse information during training. This project is developing algorithms in the SVM family that allow extra information to be used effectively during training, with the understanding that this extra information will not be available during actual operation. Examples of extra information include structural homologies between proteins in a system designed to predict structure from amino acid sequences; and values for a financial time series between the time where a prediction is made and the time of the value being predicted. Preliminary testing has shown that such extra information can dramatically reduce prediction error in the learned system compared with current generation machine learning methods that cannot use this extra information.
This project encompasses analytic research to establish performance bounds on our new algorithms, and to explore the relationships of this work to human learning. The project also includes experimental work, including construction of novel training and testing datasets; software implementation of the algorithms; and training, testing and analysis of experimental results. Areas of application include handwritten character recognition; 3-D protein structure prediction; non-linear time series prediction, for example of financial time series; and prediction of likelihood of hospital readmittance for elderly patients. This project aims to give greater insight into the nature of learning, whether in humans or machines, and seeks to formally take into account data that is today seen as only peripheral to the learning task, and impossible for current machine learning algorithms to use.
The project will produce technical articles, a book, and teaching materials explaining this research. In addition the project will produce sharable software that implements the best version of the algorithm devised during the life of the project.
|
1 |