Research Questions

An evolving list of open questions I am genuinely pursuing — at the interface of neuroscience and AI.

A short, dated list of the questions I currently find most worth chasing. Pinned not as polished claims but as active threads — what I am thinking about, where I would welcome collaboration, and what I might be wrong about. The list is updated as understanding shifts.

Last updated: 2026-04-25.


Open questions

1. How does context reshape perception and decision under similar inputs?

  • Why it matters: Across rodent empathy, LLM emotion reasoning, and conditional visual similarity (CLAY), the same input is interpreted differently when relational/temporal/social context shifts. A general account of context-conditioning seems within reach but lacks a unified formalism.
  • Open: Is “context” best modeled as a slow-moving prior, a separate routing variable, or a learned modulation gate? Do the same mechanisms appear across modalities?

2. Can neural-data-derived inductive biases meaningfully improve AI architectures — or is this just narrative?

  • Why it matters: Brain-inspired AI is often invoked, rarely tested rigorously. I want to identify cases where neuroscience-derived constraints (sparsity, predictive coding, hippocampal indexing) yield a measurable, reproducible advantage versus matched non-bio-inspired baselines.
  • Open: What is the smallest claim of “biologically informed” that is empirically defensible?

3. What does an honest evaluation of animal behavior look like with 3D + foundation models?

  • Why it matters: 3D reconstruction (3DGS, MVG) + foundation-model features promise quantitative behavioral phenotyping. But replicability across labs, species, and lighting conditions is shaky. The field needs benchmarks, not benchmarks of benchmarks.
  • Open: What set of held-out conditions would convince a skeptical neuroscientist that a behavior pipeline generalizes?

4. Where does emotion-aware LLM reasoning fail in ways that matter for downstream decisions?

  • Why it matters: Recent prompting/few-shot/ensembling improvements (e.g., my Zenodo 2025 preprint) raise emotion-recognition scores. But many failures cluster at decision-relevant edge cases (sarcasm, cultural nuance, suppression). Score-level improvements may obscure the failure shape.
  • Open: What is the right error analysis taxonomy when the downstream user is an applied agent, not a leaderboard?

5. Can the “Virtuous Cycle of NeuroAI” be operationalized with a concrete protocol, not a slogan?

  • Why it matters: Bridging biological and artificial intelligence is broadly invoked. The actual loop — data from brains → constraints on models → models → predictions about brains — needs a worked-out instance with falsifiable steps.
  • Open: Pick one cognitive function (e.g., context-dependent fear extinction, or perceptual similarity modulation) and build the closed loop end-to-end. What breaks?

6. What is the minimal setup for trustworthy cross-species behavioral comparison?

  • Why it matters: Rodent → primate → human comparisons are foundational to neuroscience, but methodological asymmetries (recording resolution, behavioral repertoire, ecological validity) muddy translation. A minimal joint protocol could change what comparisons are even meaningful.
  • Open: What is the smallest invariant we should require before claiming a behavioral phenotype “translates”?

Updates

  • 2026-04-25 — Initial registry posted.

Notes for visitors

If any of these questions overlaps with what you are working on, I would love to hear from you. Reach: biasdrive@gmail.com.

If a question is missing or seems poorly framed, that is genuinely useful feedback — I expect this list to change.