NYU PSYCH-GA 3405.004 / DS-GA 1016.003
This project is maintained by brendenlake
Here is a list of final project ideas organized by topic. For many of these ideas, a first-rate project would be a novel contribution to research in computational cognitive modeling. In the list below, each bullet point is a separate project idea.
References:
McClelland, J. L. (1981). Retrieving general and specific information from stored knowledge of specifics. In Proceedings of the third annual meeting of the cognitive science society.
References:
McClelland, J. L. (1981). Retrieving general and specific information from stored knowledge of specifics. In Proceedings of the third annual meeting of the cognitive science society.
McClelland, J. L. (2013). Integrating probabilistic models of perception and interactive neural networks: a historical and tutorial review. Frontiers in psychology, 4, 503.
References
McClelland, J. L., & Rogers, T. T. (2003). The parallel distributed processing approach to semantic cognition. Nature Reviews Neuroscience, 4(4), 310.
Large-scale learning of lexical classes. Can you discover lexical classes with a large-scale recurrent neural network (RNN)? Train a more contemporary recurrent neural network architecture (such as a LSTM) on a next word prediction task given a sizeable corpus of text, including thousands of sentences. Can you replicate some of the results from Elman (1990) on a bigger corpus, especially the hierarchical clustering results for discovering lexical categories (Lecture 3 Slide 21)?
Exploring lexical and grammatical structure in BERT or GPT2. What do powerful pre-trained models learn about lexical and grammatical structure? Explore the learned representations of a state-of-the-art language model (BERT, GPT2, etc.) in systematic ways, and discuss the learned representation in relation to how children may acquire this structure through learning.
References
Elman, J. L. (1990). Finding structure in time. Cognitive science, 14(2), 179-211.
Devlin, J., Chang, M. W., Lee, K., & Toutanova, K. (2018). Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
Radford et al., (2019). Language Models are Unsupervised Multitask Learners. arXiv preprint.
References
Gerstenberg, T., & Goodman, N. (2012). Ping Pong in Church: Productive use of concepts in human probabilistic inference. In Proceedings of the Annual Meeting of the Cognitive Science Society.
References
Feldman, N. H., & Griffiths, T. L. (2007). A rational account of the perceptual magnet effect. In Proceedings of the Annual Meeting of the Cognitive Science Society. (http://ling.umd.edu/~nhf/papers/PerceptualMagnet.pdf)
References
The OpenAI Gym: https://gym.openai.com/envs/#atari
Dubey, R., Agrawal, P., Pathak, D., Griffiths, T. L., & Efros, A. A. (preprint). Investigating human priors for playing video games. In Proceedings of the 35th International Conference on Machine Learning (ICML 2018). https://rach0012.github.io/humanRL_website/ (paper and project website)
References
The Choice Prediction Challenge website: https://cpc-18.com
Plonsky, Ori and Erev, Ido and Hazan, Tamir and Tennenholtz, Moshe, Psychological Forest: Predicting Human Behavior (May 19, 2016). Available at SSRN: https://ssrn.com/abstract=2816450 or http://dx.doi.org/10.2139/ssrn.2816450
Joshua Peterson, David Bourgin, Mayank Agrawal, Daniel Reichman, Thomas Griffiths (2021). Using large-scale experiments and machine learning to discover theories of human decision-making. Science. https://www.science.org/doi/abs/10.1126/science.abe2629
Contribute to open science! As mentioned in the lecture on categorization there are a variety of different theories of how people learn categories and concepts from examples, and many of these models draw from similar approaches to machine classification. Currently there is a community-led effort to implement all existing category learning models in R so they can be simultaneously compared to the same sets of human data patterns. While the project is well developed and documented there are many example models which have not yet been implement (some of which are actually somewhat easy). One nice final project would be to read one of the papers listed below which describe a famous formal model of human categorization and implement it for submission to the catlearn R package. This is best for a group with some expertise in R (as opposed to python). If your group makes a report showing how this model does on the existing data set in catlearn it would make a nice final paper and by making a pull request against the catlearn package your work for class might live forever to help advance science! You could also choose to implement these models in python in which case you could still make an impact by helping to verify the previous results (see below).
References:
Catlearn R pacakage: https://ajwills72.github.io/catlearn/
Examples:
RULEX is a simple type of decision tree algorithm - Nosofsky, R. M., Palmeri, T. J., & McKinley, S. C. (1994). Rule-plus-exception model of classification learning. Psychological Review, 101(1), 53-79.here
ATRIUM is a hybrid rule and nearest neighbor/exemplar algorithm that is trained used backpropogation and gradient descent - Erickson, M. A., & Kruschke, J. K. (1998). Rules and exemplars in category learning. Journal of Experimental Psychology: General, 127(2), 107-140. here
Psychology is largely an empirical science and thus findings need to be independently replicated before they should be widely accepted. This is true for computational cognitive modeling as well. A line of interesting final projects would be to pick a recent computational modeling paper and to re-implement and verify the results reported by the authors. In doing this exercise you might come up with your own ideas about a feature to alter or change in their simulations that might be interesting. A couple good sources for papers would be to look at the titles from the most recent Proceedings of the Cognitive Science society or a new journal called Computational Brain and Behavior.
References