{"id":5450,"date":"2026-03-13T19:38:49","date_gmt":"2026-03-13T19:38:49","guid":{"rendered":"https:\/\/sinatootoonian.com\/?p=5450"},"modified":"2026-03-22T12:30:57","modified_gmt":"2026-03-22T12:30:57","slug":"neurips-2025-day-3","status":"publish","type":"post","link":"https:\/\/sinatootoonian.com\/index.php\/2026\/03\/13\/neurips-2025-day-3\/","title":{"rendered":"Neurips 2025 Day 3"},"content":{"rendered":"\n<p>My notes on Day 3 of NeurIPS 2025. Wanted to wait till when they were more polished, but after three months I&#8217;m just posting as is.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Poster Session 3<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\"><a href=\"https:\/\/neurips.cc\/virtual\/2025\/loc\/san-diego\/poster\/117029\">Seemingly Redundant Modules Enhance Robust Odor Learning in Fruit Flies<\/a><\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Asks whether spike-frequency adaptation and lateral inhibition in the fly olfactory system are redundant or complementary<\/li>\n\n\n\n<li>Tests a baseline model (no SFA or LI), against models with just one, and the full model<\/li>\n<\/ul>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"620\" src=\"https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-4-1024x620.png\" alt=\"\" class=\"wp-image-5451\" srcset=\"https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-4-1024x620.png 1024w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-4-300x182.png 300w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-4-768x465.png 768w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-4-1536x931.png 1536w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-4-2048x1241.png 2048w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<ul class=\"wp-block-list\">\n<li>SFA and LI are better than each other depending on injected noise.<\/li>\n\n\n\n<li>Full model performs best across all noise conditions:<\/li>\n<\/ul>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"365\" src=\"https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-5-1024x365.png\" alt=\"\" class=\"wp-image-5452\" srcset=\"https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-5-1024x365.png 1024w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-5-300x107.png 300w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-5-768x274.png 768w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-5-1536x547.png 1536w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-5-2048x729.png 2048w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<h3 class=\"wp-block-heading\"><a href=\"https:\/\/neurips.cc\/virtual\/2025\/loc\/san-diego\/poster\/117502\">Neurons as Detectors of Coherent Sets in Sensory Dynamics<\/a><\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Coherent sets: subsets of state-space that move together in time.<\/li>\n\n\n\n<li>Neurons, brains want to find representations that track such states.<\/li>\n\n\n\n<li>Koopman operator: Linear operator in high-dimensional space that reports future state of an observable.\n<ul class=\"wp-block-list\">\n<li>Linearity means we can write it as $$K_\\tau =  \\sum_{i=0} ^\\infty  e^{\\lambda_i \\tau} u_i v_i^T.$$  <\/li>\n<\/ul>\n<\/li>\n\n\n\n<li>Maximize P(f(X(t+\\tau)) \\in B | f(X(t)) \\in A) + P(f(X(t+\\tau)) \\in B^c | f(X(t)) \\in A^c)<\/li>\n\n\n\n<li>Use the observable that  <\/li>\n\n\n\n<li>TBC<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\"><a href=\"https:\/\/neurips.cc\/virtual\/2025\/loc\/san-diego\/poster\/116813\">BrainFlow: A Holistic Pathway of Dynamic Neural System on Manifold<\/a><\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>The problem: linking structural connectivity and functional connectivity.<\/li>\n\n\n\n<li>Both connectivities live on the manifold of symmetric, positive-definite matrices (SPD).<\/li>\n\n\n\n<li>Can link them using flows along geodesics on the manifold.<\/li>\n\n\n\n<li>The same structural connectivity will lead to different functional connectivities during different tasks.<\/li>\n\n\n\n<li>Conversely, functional connectivity produced in different tasks should lead back to the same structural connectivity.<\/li>\n\n\n\n<li>Problem: Computing flows on the SPD manifold is computationally expensive.<\/li>\n\n\n\n<li>PSD manifold can be linked to the manifold of lower triangular matrices by the Cholesky decomposition.<\/li>\n\n\n\n<li>Recent work  shows computation on the Cholesky manifold is much easier.<\/li>\n\n\n\n<li>Solution: Leveraged this recent work to compute flows linking structure and function on the cholesky manifold.<\/li>\n<\/ul>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"482\" src=\"https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-6-1024x482.png\" alt=\"\" class=\"wp-image-5455\" srcset=\"https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-6-1024x482.png 1024w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-6-300x141.png 300w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-6-768x362.png 768w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-6.png 1422w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<ul class=\"wp-block-list\">\n<li>To get flows from different FC back to the same SC, they add a term that penalizes differences in the Cholesky manifold states.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\"><a href=\"https:\/\/neurips.cc\/virtual\/2025\/loc\/san-diego\/poster\/116784\">Dimensionality Mismatch Between Brains and Artificial Neural Networks<\/a><\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Measured dimensionality of representations across layers in both humans (fMRI data) and ANNs.<\/li>\n\n\n\n<li>Use both linear dimensionality (participation ratio) and non-linear dimensionality, defined as below.\n<ul class=\"wp-block-list\">\n<li>As a side note, see Mackay and Gharamani&#8217;s <a href=\"http:\/\/www.inference.org.uk\/mackay\/dimension\/\">note<\/a> on this estimator.<\/li>\n<\/ul>\n<\/li>\n<\/ul>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"218\" src=\"https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-7-1024x218.png\" alt=\"\" class=\"wp-image-5460\" srcset=\"https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-7-1024x218.png 1024w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-7-300x64.png 300w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-7-768x164.png 768w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-7-1536x327.png 1536w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-7.png 1766w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<ul class=\"wp-block-list\">\n<li>The authors found that in humans both measures increase as one moves from V1 towards ventral stream. <\/li>\n\n\n\n<li>In ANNs, intrinsic dimensionality seems to peak at intermediate layers. <\/li>\n<\/ul>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"569\" src=\"https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-10-1024x569.png\" alt=\"\" class=\"wp-image-5463\" srcset=\"https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-10-1024x569.png 1024w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-10-300x167.png 300w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-10-768x427.png 768w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-10-1536x854.png 1536w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-10-2048x1139.png 2048w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Suggests that deep-layers in ANN may be compressing their representations, losing abstraction ability.\n<ul class=\"wp-block-list\">\n<li>Though interestingly the peak still happens at a higher value than in humans.<\/li>\n<\/ul>\n<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\"><a href=\"https:\/\/neurips.cc\/virtual\/2025\/loc\/san-diego\/poster\/115221\">Modeling Neural Activity with Conditionally Linear Dynamical Systems<\/a><\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Model nonlinear dynamics as linear conditioned on inputs:<\/li>\n<\/ul>\n\n\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"856\" height=\"188\" src=\"https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-11.png\" alt=\"\" class=\"wp-image-5466\" srcset=\"https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-11.png 856w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-11-300x66.png 300w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-11-768x169.png 768w\" sizes=\"auto, (max-width: 856px) 100vw, 856px\" \/><\/figure>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Notice how inputs modify the parameters, instead of coming in as the usual additive $Bu_t$ terms.<\/li>\n\n\n\n<li>GP priors on each element of $\\mathbf{A, b, C, d}$.<\/li>\n\n\n\n<li>Get all the benefits of linear systems for inference, prediction etc.<\/li>\n\n\n\n<li>Can learn parameters with EM: \n<ul class=\"wp-block-list\">\n<li>E-step: Use current parameter estimates to infer latent states.<\/li>\n\n\n\n<li>M-step: Solve for parameters by finding coefficients in fixed GP basis.\n<ul class=\"wp-block-list\">\n<li>Closed form when assuming Gaussian noise.<\/li>\n<\/ul>\n<\/li>\n<\/ul>\n<\/li>\n\n\n\n<li>Can be seen as conditions specific linearizations of the dynamics.<\/li>\n\n\n\n<li>Temporal correlation of GP prior can be tuned to capture interesting limits:\n<ul class=\"wp-block-list\">\n<li>When time constants are small, learns independent, per-condition dynamics.<\/li>\n\n\n\n<li>When time constants are infinite, fits a fixed, time-invariant linear dynamics (the classical case).<\/li>\n<\/ul>\n<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\"><a href=\"https:\/\/neurips.cc\/virtual\/2025\/loc\/san-diego\/poster\/119953\">Connecting Neural Models Latent Geometries with Relative Geodesic Representations<\/a><\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Neural networks (e.g. autoencoders) trained on different samples of the same data often learn similar internal representations.<\/li>\n\n\n\n<li>These can often  be aligned with simple linear transformations.<\/li>\n\n\n\n<li>Suggests that the same underlying manifold is being parameterized.<\/li>\n\n\n\n<li>One canonicalization is to not represent the latents in absolute terms, but in relative terms using distances to certain anchor points.<\/li>\n\n\n\n<li>Previously this has been done using cosine similarity.<\/li>\n\n\n\n<li>This work: the decoder implicitly defines a pullback metric on the latent space.<\/li>\n\n\n\n<li>Instead of cosine distance, use geodesic distance defined by this metric.<\/li>\n<\/ul>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"418\" src=\"https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-12-1024x418.png\" alt=\"\" class=\"wp-image-5469\" srcset=\"https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-12-1024x418.png 1024w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-12-300x122.png 300w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-12-768x313.png 768w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-12-1536x626.png 1536w, https:\/\/sinatootoonian.com\/wp-content\/uploads\/2025\/12\/image-12-2048x835.png 2048w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<p><\/p>\n","protected":false},"excerpt":{"rendered":"<p>My notes on Day 3 of NeurIPS 2025, posting three months later as is.<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"footnotes":""},"categories":[1,167],"tags":[],"class_list":["post-5450","post","type-post","status-publish","format-standard","hentry","category-blog","category-conference"],"acf":[],"_links":{"self":[{"href":"https:\/\/sinatootoonian.com\/index.php\/wp-json\/wp\/v2\/posts\/5450","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/sinatootoonian.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/sinatootoonian.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/sinatootoonian.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/sinatootoonian.com\/index.php\/wp-json\/wp\/v2\/comments?post=5450"}],"version-history":[{"count":12,"href":"https:\/\/sinatootoonian.com\/index.php\/wp-json\/wp\/v2\/posts\/5450\/revisions"}],"predecessor-version":[{"id":7311,"href":"https:\/\/sinatootoonian.com\/index.php\/wp-json\/wp\/v2\/posts\/5450\/revisions\/7311"}],"wp:attachment":[{"href":"https:\/\/sinatootoonian.com\/index.php\/wp-json\/wp\/v2\/media?parent=5450"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/sinatootoonian.com\/index.php\/wp-json\/wp\/v2\/categories?post=5450"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/sinatootoonian.com\/index.php\/wp-json\/wp\/v2\/tags?post=5450"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}