tag:blogger.com,1999:blog-5694696150342161606.post5398991406468174541..comments2014-04-28T08:18:28.717-07:00Comments on Visual Neurons and Machines: scenes objects based repAnonymoushttp://www.blogger.com/profile/05700100167215401132noreply@blogger.comBlogger32125tag:blogger.com,1999:blog-5694696150342161606.post-58921737953936766422014-04-02T11:06:36.694-07:002014-04-02T11:06:36.694-07:00One thing that struck me after class was that the ...One thing that struck me after class was that the PPA was shown to be more correlated with LOC than RSC, so that kind of goes against the hypothesis that PPA is basically latching on to the 3D structure or spatial information from the furniture objects. PPA is also encoding the object content as well. But again all this depends on how fair it is to use the correlation metric for comparison between regions of the brain, how long is the time to process the images etc.Anonymoushttps://www.blogger.com/profile/08846162704066452969noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-41058407990202908882014-04-02T08:56:28.733-07:002014-04-02T08:56:28.733-07:00I would be curious to know people's opinions o...I would be curious to know people's opinions on what this all means for the visual process in general. I.E. if we forget about PPA/SRC/LOC for a minute and simply make the claim: "Spatial information about scenes (open/closed) is processed in one area of the brain, object identity in another area and object pose in another area" what does that actually tell us about the visual problem? What would be different if they were all processed as part of one big block of functionality? To me it makes obvious sense that they would be separate: the individual pieces of data seem to be relatively uncorrelated, and seem to require different process to evaluate, so it would make sense to have dedicated regions. Does anyone have any other thoughts on what we might get from this information?Anonymoushttps://www.blogger.com/profile/09988915607928653106noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-58490892942755575102014-04-02T08:45:10.178-07:002014-04-02T08:45:10.178-07:00I personally like the use of synthetic data. It&#...I personally like the use of synthetic data. It's important not to extrapolate the conclusions too far, but there is something to separating out the important things we want to test for... removing clutter if you will. It might have been more compelling though if they simply constructed real physical scenes that were very simple and photographed them, rather than using CG models. Honestly if you're going through all the trouble of running fMRI on 20 patients (at hundreds of dollars an hour), you think they could have found a bare room and an old couch somewhere and simply taken a bunch of photos. Also I agree that they probably should have done more comparisons against complex scenes.Anonymoushttps://www.blogger.com/profile/09988915607928653106noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-8412144138515096332014-04-02T08:05:27.716-07:002014-04-02T08:05:27.716-07:00I'm actually not sure how much the PPA is able...I'm actually not sure how much the PPA is able to decode. Last class I defended this correlation approach as being better than an SVM. But one problem I now see is that I really don't know what 0.02 decoding here means. Is this strong decoding? Yes, yes, it's statistically significant (i.e., there's something going on), but the question I have is whether this can be explained by other stuff going on in the brain that's not really "semantic-object" tuned. Basically: is this decoding accuracy good enough that it's not just low-level statistics or some other stuff floating around the PPA.Anonymoushttps://www.blogger.com/profile/06191884207078168062noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-44761935282238462602014-04-02T07:55:52.677-07:002014-04-02T07:55:52.677-07:00Till now I still do not get a clear picture with r...Till now I still do not get a clear picture with respect to the role of the corresponding regions in the brain for scene understanding, especially PPA. Different papers we read use different experimental setup and dataset, and hence sometimes lead to controversial conclusions. I thinks this is probably because of different methodology between computer vision and neuroscience. In computer vision, we are accustomed to proving our assumption using extensive experiments on different dataset and even for different vision tasks. Without doing so, our paper would have high chance of being rejected. Meanwhile, for neuroscience paper, probably partially due to the hardness of conducting comprehensive experiments, they are usually simplified, such as simplified dataset and simplified parameter tuning procedure. Hence, the conclusions are usually skeptical for vision guys. Yuxiong Wanghttps://www.blogger.com/profile/07729560385478615651noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-74649448276960790172014-04-02T07:50:41.316-07:002014-04-02T07:50:41.316-07:00I agree with their philosophy of disentangling and...I agree with their philosophy of disentangling and controlling different factors by using simplified synthetic data due to the indeed complexity of the issue. However, i think their conclusions conducted by the usage of only synthetic data is unguaranteed. Simply combining both synthetic and natural data, i.e., repeating the experiments using both data, is also not enough. A better approach is to use the gradual evolution of data. For example, start from some simplified synthetic data, and then gradually add other objects, backgrounds and geometric layout to make the scene become more complex. An alternative is that start from some complex scene image, and then gradually discard some objects and backgrounds to simplify the scene. In both scenarios, test whether and how the response of the corresponding areas in the brain change. Yuxiong Wanghttps://www.blogger.com/profile/07729560385478615651noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-44578939844257118652014-04-02T06:13:34.915-07:002014-04-02T06:13:34.915-07:00The paper emphasizes the connections from LOC and ...The paper emphasizes the connections from LOC and RSC into PPA and how PPA receives information from several brain regions (hippocampus, V4 etc.).<br />What does PPA output too?<br /><br />I'm interested in this question because then we may hypothesize what the brain does with the scene information.M Aravindhhttps://www.blogger.com/profile/06228466877204452065noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-68803504052601892732014-04-02T06:11:03.135-07:002014-04-02T06:11:03.135-07:00I agree that synthetic data has its flaws, but don...I agree that synthetic data has its flaws, but don't you think it's important to understand which changes in a scene matter the most? By making the data as simple as possible, they can see the effect of the slightest change (desk vs bed). I think this is more difficult with messy real world data. Jacob Walkerhttps://www.blogger.com/profile/15006407210819891433noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-71067781950516574162014-04-02T06:08:03.914-07:002014-04-02T06:08:03.914-07:00I think it's important think that biological v...I think it's important think that biological vision is a means to an end. Organisms want to know where they are, find food, etc. PPA, LOC, and RSA may all be toward different cognitive ends. I agree that the data is contrived, but it may make sense for PPA to be sensitive to objects as well if it is indeed part of a larger pathway that aids in navigation. It may be a way of noting "landmarks"; I know that this scene (or I) was at place X and it had an object Y. Jacob Walkerhttps://www.blogger.com/profile/15006407210819891433noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-42816917818975963402014-04-02T06:07:52.138-07:002014-04-02T06:07:52.138-07:00If PPA is able to decode between semi-fine grained...If PPA is able to decode between semi-fine grained furniture categories then its definitely not latching onto simple cues like perspective.M Aravindhhttps://www.blogger.com/profile/06228466877204452065noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-85966344146894117752014-04-02T05:05:32.260-07:002014-04-02T05:05:32.260-07:00I think we need results from both kinds of data - ...I think we need results from both kinds of data - synthetic and natural. The authors argue, in the introduction, that naturalistic photographic stimuli has led to conflicting results. In such a situation a synthetic data based experiment seems very useful.<br /><br />I also feel that the brain areas discussed here are not so abstract as to have different pathways for synthetic vs real scenes - just a guess.M Aravindhhttps://www.blogger.com/profile/06228466877204452065noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-19734964221988672692014-04-02T02:23:27.814-07:002014-04-02T02:23:27.814-07:00I wonder if there have been tries in vision, where...I wonder if there have been tries in vision, where people have used different methods on the same data and brought them together to add more to the over-all semantics. Like, one analysis for say high-level recognition, another method for fine-grained analysis and finally bringing them together to give a consolidated result. Can any one mention some of such experiments? If any?Priyamhttps://www.blogger.com/profile/15315859320880746009noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-64384746193744392322014-04-02T01:41:19.795-07:002014-04-02T01:41:19.795-07:00Taking inspiration PPA's use of both what and ...Taking inspiration PPA's use of both what and where -<br />Maybe in computer vision we can use cues from object recognition to do spatial understanding eg. identifying part of image as sky means it's far away always. This reminds me of a CVPR paper that used semantic labels to inform depth prediction - (http://users.cecs.anu.edu.au/~sgould/papers/cvpr10-depth.pdf) Gauravhttps://www.blogger.com/profile/18207527108646043527noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-41011982768610443572014-04-02T01:40:12.376-07:002014-04-02T01:40:12.376-07:00Reading this paper I was immediately reminded of t...Reading this paper I was immediately reminded of the Computer vision's GIST feature - http://cvcl.mit.edu/papers/IJCV01-Oliva-Torralba.pdf <br />PPA does what and where' processing and maybe this gives the inspiration for a general scene understanding feature like GIST ? <br />RSC is more related to the where aspect - might be useful for actual robotic applications of fast motion through environments. Read replicating driving/flying on robots.Gauravhttps://www.blogger.com/profile/18207527108646043527noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-82505440190965236422014-04-01T23:44:00.116-07:002014-04-01T23:44:00.116-07:00This paper investigated synthetic scenes with one ...This paper investigated synthetic scenes with one object, and it seems to be suggesting that there are 2 streams of processing, one for spacial properties, and the other for object properties. The relationship between different objects in the scene seems to be important. Object properties may be used as cues to spacial properties. For example we might rely on the position of objects on a smooth surface to deduce where that surface is. It would be interesting to see where and how the 2 streams of processing merge to form a consistent representation. Anonymoushttps://www.blogger.com/profile/02936154152799787515noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-82065584167506409172014-04-01T23:42:38.796-07:002014-04-01T23:42:38.796-07:00Keeping with the theme of simmering discontentment...Keeping with the theme of simmering discontentment in this thread, in the discussions section when they describe their general approach, they claim that they use their formulated framework to "inform the design of complementary studies that use both *naturalistic* and artificial scenes to understand the nature of information being represented in the scene network."<br /><br />Are they referring to the monkey paper? Isn't that too big a jump? Or am I just parsing this incorrectly?Kumar Shauryahttps://www.blogger.com/profile/14347242113456802229noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-50822325510297028722014-04-01T23:17:19.830-07:002014-04-01T23:17:19.830-07:00To add to the skepticism, their take-away message ...To add to the skepticism, their take-away message is much too broad for the experiment they actually did. I'm particularly peeved by the idea that object information is represented by invariance to backgrounds. Simply low-level features could get you the same thing.Alliehttps://www.blogger.com/profile/12551902713568107756noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-45656558589008012732014-04-01T23:14:51.057-07:002014-04-01T23:14:51.057-07:00I am similarly skeptical about the synthetic image...I am similarly skeptical about the synthetic images. They used gray backgrounds but color objects, and it is imaginable how this might not be a good idea particularly in the case of "open" scenes.Anonymoushttps://www.blogger.com/profile/02936154152799787515noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-13045753226474806762014-04-01T22:59:34.705-07:002014-04-01T22:59:34.705-07:00Interesting the PPA has also been shown to capture...Interesting the PPA has also been shown to capture global statistics and textures. Following a discussion from the last class on dividing the scene recognition task into structure, content and style, would it be fair to say then that according to this paper, PPA is sensitive to all three (structure, content, style)? It seems to suggest that PPA is integrating inherited information from different regions of the brain to culminate the process of scene understanding. Anonymoushttps://www.blogger.com/profile/08846162704066452969noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-42319266606433107862014-04-01T22:57:43.161-07:002014-04-01T22:57:43.161-07:00The author mentioned that hey are using furniture ...The author mentioned that hey are using furniture because they can realistically embedded in a spatial environment. But many other flat objects like flat carpet, portrait etc can be easily embedded in a spatial environment and can be used in their experiment.Anonymoushttps://www.blogger.com/profile/12616703658448712877noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-85867342704422415602014-04-01T22:47:29.653-07:002014-04-01T22:47:29.653-07:00I liked the way in which author has shown that LOC...I liked the way in which author has shown that LOC is responsible for processing, RSC for scene layout and PPA for both. Also at last they have shown RSC is correlated with PPA which is also correlated with LOC. So is this correlation is enough to imply actual physical connection between these regions or exchange of information between them?Anonymoushttps://www.blogger.com/profile/12616703658448712877noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-82063587463210049512014-04-01T22:46:56.587-07:002014-04-01T22:46:56.587-07:00The authors also mention in the 'Discussion...The authors also mention in the 'Discussion' section that the objects that they use have navigational affordances which provide diagnostic details about spatial environment. So ultimately PPA may be sensitive to spatial information alone from different sources. Other objects such as manipulable tools or complex backgrounds may not lead to similar results.Anonymoushttps://www.blogger.com/profile/08846162704066452969noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-35407681357630458502014-04-01T22:39:53.490-07:002014-04-01T22:39:53.490-07:00On the other hand, these big objects also make the...On the other hand, these big objects also make the case for the LOC results stronger.Ishanhttps://www.blogger.com/profile/14430904277632185537noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-16156336544888085202014-04-01T22:36:09.579-07:002014-04-01T22:36:09.579-07:00This paper seems to reinforce the way vision is pe...This paper seems to reinforce the way vision is perceived in modern times. The object recognition sub-field, the geometry (layout, single image 3D) sub-field having not too much cross talk.<br />I am highly disturbed by the fact that they use furniture objects (which themselves give strong cues of box-like 3D information), and nothing else (which may include more 2D/flat objects). The fact that we are using such objects might be a reason why the PPA gives a really high response for objects. It might just be latching on to the "perspectiveness" of these objects.Ishanhttps://www.blogger.com/profile/14430904277632185537noreply@blogger.comtag:blogger.com,1999:blog-5694696150342161606.post-5342467298636866402014-04-01T22:18:12.735-07:002014-04-01T22:18:12.735-07:00To add to your second point, the simplicity of sce...To add to your second point, the simplicity of scenes might have some role to play in PPA processing objects in these experiments...Abhinav Shrivastavahttps://www.blogger.com/profile/04212112791106153178noreply@blogger.com