My research spans several areas of vision research with common threads connecting the different topics. I study the perception of shape, both 2D shapes and 3D surfaces, and saccadic eye movements. I have a background in both Psychology (Ph.D.) and Computer Science (M.S.), and my work in both areas has involved computational modeling and psychophysics. In all the areas of my research one of the goals has been uncovering what information is available to the observer, how that information is represented (frequently looking at whether the local or global information is encoded), and whether or not the representation used can be flexible and change with changing task demands.
      I am currently working as a Postdoctoral Fellow in Computer Science under the direction of Prof. Sven Dickinson. We are investigating using the junctions of superpixels to improve image segmentation and the interpretation of the figure/ground organization of the scene.
My previous postdoctoral position was in the lab of Prof. Richard Murray. I studied how the human visual system represents scene lighting information and how the visual system interprets shape from shading. Most models of shape from shading make a global lighting assumption, using a single fixed light source. Our work suggests that the human visual system does not make a global lighting direction assumption. Instead lighting direction is represented locally, and the area of the local estimate is related to the frequency of shape changes. The most surprising result of the project is that observers cannot always consciously report the local lighting direction, but their visual system still uses the appropriate local lighting information when interpreting the shape of an object. Subjects were shown a surface where one quadrant was lit from one lighting direction and the other 3 quadrants were lit from a lighting direction with a difference in tilt of 90 degrees. Subjects were unable to identify the oddly lit region, even though estimations of shape from shading were just as accurate in the odd region as in the consistently lit regions. This result suggests that the visual system uses the appropriate local lighting information when estimating shape from shading even when that information is not reportable by the subjects.
I have collaborated with Prof. James Elder and Dr. Ingo Fründ to investigate 2D shape preception, and develop a classification image paradigm for that can be used to efficiently probe the internal representation used by the human visual system. In these experiments subjects classified a shape into one of two categories. We designed the experiment in such a way that in a discrimination task all frequencies of the shape were equally useful, but the data showed that humans were unable to make use of high frequency shape information. Our estimated classification images were better predictors of human performance than making an assumption that humans perform as the ideal observer with internal noise, suggesting that humans pay more attention to the more global structure of a shape than to the fine details when performing our task, even though all frequencies had the same amount of information. This is consistent with my other shape work, stressing the importance of global and regional representations.
I earned my PhD working in the Visual Cognition Lab under the direction of Dr. Jacob Feldman
Many of my projects focus on investigating whether visual representations encode local/detailed information or global information. In Wilder, Feldman, and Singh (2011), we created a model for shape classification that could accurately predict a human’s classification of a novel shape. Using a skeletal representation of shapes (Feldman & Singh, 2006) we created a naïve Bayesian classifier for leaves and animals. By morphing together animals and leaves we created novel shapes. These novel shapes were presented to both human observers and the classifier. The classifications made by the human observers could be accurately predicted by the naïve Bayesian classifier. This shows that a skeletal representation contains the necessary information for accurate classification, possibly due to the skeletal representation capturing relevant information about part structure, and that classifiers based upon the shape skeleton can be tuned to the statistics of natural shapes.
      Also, while at Rutgers, I collaborated with Prof. Eileen Kowler on eye movements. In Wilder, et al. (2009) we measured how much attentional cost there is to making an eye movement. In another project, we studied how an observer decides when to move the eye to a new location. Standard models of information accumulation model the process stochastically, where the information accumulation is simulated with no direct connection to an actual stimulus. We designed a stimulus to allow for the direct measurement of information at every point in time. Subjects saw a display where dots appeared slowly, one at a time. Subjects were required to decide if the mean of the distribution from which these dots are sampled was to the right or left of a reference line. Subjects let dots accumulate until they either decided they had waited long enough or had enough information to make the decision about the location of the mean of the dot generating distribution. One key part of this project was the creation of a task that allowed us to give the observer relevant information in measurable chunks, giving us an upper bound on the amount of information used by the observer to complete the task. We directly compared predictions from an information accumulation model and an internal clock (timer) model. Neither model, nor standard variations on those models, was able to completely account for our subjects’ behavior. We find that a hybrid evidence/timer model is necessary to account for subject behavior.