Advertisement · 728 × 90

Posts by Taku Ito

Preview
The fall of the theorem economy How AI could destroy mathematics and barely touch it

A long read about the state of AI and mathematics.

davidbessis.substack.com/p/the-fall-o...

1 day ago 9 6 1 3
Video

🧵 New preprint led by @bingbrunton.bsky.social, @elliottabe.bsky.social, @lawrencehu.bsky.social

We gave a worm brain control of a fly body and it walked

What did we learn? Nothing, other than deep reinforcement learning is effective

We call it the digital sphinx

www.biorxiv.org/content/10.6...

4 weeks ago 397 147 9 27
Video

www.percepta.ai/blog/can-llm...

As a research lark at Percepta, Christos embedded a computer into an LLM, showed that it could solve the hardest Sudokus, and then as a side bonus built an exponentially faster attention

1 month ago 314 58 21 43

Bullshit Bench V2

new: 100 questions across several domains

- Anthropic & Qwen still on top
- Reasoning seems to hurt
- New models are *not* better than old (except Claude)
- Seems to be independent of domain

github.com/petergpt/bul...

1 month ago 105 10 9 3
Preview
Text-to-LoRA: Instant Transformer Adaption While Foundation Models provide a general tool for rapid content creation, they regularly require task-specific adaptation. Traditionally, this exercise involves careful curation of datasets and repea...

Sakana has developed a way to, if I understand correctly, instantly generate LORAs on demand from long texts or documents

arxiv.org/abs/2506.06105
arxiv.org/abs/2602.15902

1 month ago 54 6 3 4
Preview
US science after a year of Trump: what has been lost and what remains A series of graphics reveals how the Trump administration has sought historic cuts to science and the research workforce.

Trump has been in office for one year. We at @nature.com did a deep dive looking at the administration's disruption of science in numbers.

Take a look—the numbers are staggering. By me, @dangaristo.bsky.social, Jeff Tollefson, @kimay.bsky.social, & help from @noamross.net @scott-delaney.bsky.social

3 months ago 504 317 10 30
This line graph illustrates the percentage change in agency staff levels from the previous year for nine major U.S. federal scientific and health organizations between the fiscal years 2016 and 2025. The agencies tracked include the CDC, Department of Energy, EPA, FDA, NASA, NIH, NIST, NOAA, and NSF. For the majority of the timeline between 2016 and 2023, the agencies show relatively stable fluctuations, generally staying within a range of +5% to -5% change per year. However, there is a dramatic and uniform plummet starting in the 2024–25 period. Every agency depicted shows a sharp downward trajectory, with staffing losses ranging from approximately -15% to over -25%. The Environmental Protection Agency (EPA) shows the most significant decline, dropping to roughly -26%, while the National Institute of Standards and Technology (NIST) shows the least severe but still substantial drop at approximately -15%.

This line graph illustrates the percentage change in agency staff levels from the previous year for nine major U.S. federal scientific and health organizations between the fiscal years 2016 and 2025. The agencies tracked include the CDC, Department of Energy, EPA, FDA, NASA, NIH, NIST, NOAA, and NSF. For the majority of the timeline between 2016 and 2023, the agencies show relatively stable fluctuations, generally staying within a range of +5% to -5% change per year. However, there is a dramatic and uniform plummet starting in the 2024–25 period. Every agency depicted shows a sharp downward trajectory, with staffing losses ranging from approximately -15% to over -25%. The Environmental Protection Agency (EPA) shows the most significant decline, dropping to roughly -26%, while the National Institute of Standards and Technology (NIST) shows the least severe but still substantial drop at approximately -15%.

This is the most astonishing graph of what the Trump regime has done to US science. They have destroyed the federal science workforce across the board. The negative impacts on Americans will be felt for generations, and the US might never be the same again.

www.nature.com/immersive/d4...

3 months ago 14834 8559 92 794

One of my favorite findings: Positional embeddings are just training wheels. They help convergence but hurt long-context generalization.

We found that if you simply delete them after pretraining and recalibrate for <1% of the original budget, you unlock massive context windows. Smarter, not harder.

3 months ago 219 33 8 1
Advertisement

Oh wow, deepseek is starting to make serious progress on LLMs that offload memory to external storage: github.com/deepseek-ai/...

3 months ago 217 26 6 8
Schematic depicting cortical-subcortical interactions during multi-task learning

Schematic depicting cortical-subcortical interactions during multi-task learning

Excited to see our paper with @mwcole.bsky.social finally out in peer-reviewed form @natcomms.nature.com! We examine how the human brain learns new tasks and optimizes representations over practice…1/n

5 months ago 24 7 1 0
Preview
AI discovers learning algorithm that outperforms those designed by humans An artificial-intelligence algorithm that discovers its own way to learn achieves state-of-the-art performance, including on some tasks it had never encountered before.

Did you know that AI can figure out its own way to learn, and that its way is better than one designed by humans? Read more in a @nature.com N&V (and the original paper is in the comment) 🧪 www.nature.com/articles/d41...

5 months ago 6 2 2 0

Our work with @pawa-pawa.bsky.social is out in Nature Machine Intelligence! The choice of activation function affects the representations, dynamics, and circuit solutions that emerge in RNNs trained on cognitive tasks. Activation matters!
www.nature.com/articles/s42...

5 months ago 42 11 0 0

(repost welcome) The Generative Model Alignment team at IBM Research is looking for next summer interns! Two candidates for two topics

🍰Reinforcement Learning environments for LLMs

🐎Speculative and non-auto regressive generation for LLMs

interested/curious? DM or email ramon.astudillo@ibm.com

6 months ago 19 14 1 1
Preview
Why I left academia and neuroscience Don't worry, this isn't yet another story of rage-quitting.

Michael X Cohen on why he left academia/neuroscience.
mikexcohen.substack.com/p/why-i-left...

6 months ago 95 36 7 14
Preview
Arousal as a universal embedding for spatiotemporal brain dynamics - Nature Reframing of arousal as a latent dynamical system can reconstruct multidimensional measurements of large-scale spatiotemporal brain dynamics on the timescale of seconds in mice.

Nature research paper: Arousal as a universal embedding for spatiotemporal brain dynamics

go.nature.com/4nMUgYz

6 months ago 31 12 0 2
Post image

Lab’s latest is out in Imaging Neuroscience, led by Kirsten Peterson: “Regularized partial correlation provides reliable functional connectivity estimates while correcting for widespread confounding”, where we demonstrate a major improvement to standard fMRI functional connectivity (correlation) 1/n

7 months ago 75 30 6 0
Advertisement
Preview
Can AI generate truly novel algorithms? A decades-old approach to measuring algorithmic complexity could provide a window into better understanding how AI systems compute.

Formalizing AI computation in terms of algorithmic complexity can offer a formal way to quantify AI systems while offering a principled foundation to build more algorithmically capable systems in the future.
Blog: research.ibm.com/blog/ai-algo...
arXiv: arxiv.org/abs/2411.05943

8 months ago 0 0 0 0

While using AI models to generate code is commonplace these days, we still do not fully understand the limits of the complexity of the code these models can formulate.
3/n

8 months ago 0 0 1 0
Video

Using circuits to formalize algorithmic problems for AI models (e.g., depth as time complexity, size as space complexity), we can quantify the complexity of circuit computations (algorithmic complexity) an AI model can perform.
2/n

8 months ago 0 0 1 0
Video

What complexity of algorithms can AI compute? In a new paper with colleagues at IBM Research, we explore how circuit complexity theory can help quantify the degree of algorithmic generalization in AI systems. www.nature.com/articles/s42...
@natmachintell.nature.com
#ML #AI #MLSky
1/n

8 months ago 17 5 1 1
Post image

Mental health research is at a turning point—breakthroughs can transform lives, but only with bold action, investment, and open collaboration. The time for action is now. Read our full statement here: childmind.org/blog/can-sci...

1 year ago 15 7 0 0
Post image

Out today in Nature Machine Intelligence!

From childhood on, people can create novel, playful, and creative goals. Models have yet to capture this ability. We propose a new way to represent goals and report a model that can generate human-like goals in a playful setting... 1/N

1 year ago 135 41 5 4

New preprint! Ziyan and I explore how task order impacts continual learning in neural networks and how to optimize it. Our analysis highlights two key principles for better task sequencing.
Check it out: arxiv.org/pdf/2502.03350

1 year ago 7 3 0 0

The entire website for the NIH Office of Research on Women's Health (ORWH) is very nearly stripped bare. This is so, so devastating. orwh.od.nih.gov/research/fun...

1 year ago 974 615 44 65
Advertisement
Preview
Discretized representations in V1 predict suboptimal orientation discrimination - Nature Communications How animals generate perceptual decisions remains poorly understood. Here, the authors show that during a discrimination task, the mouse visual cortex does not encode the orientations of the cues but ...

New paper out! 🚨 📰 With @batuhanerkat.bsky.social, John McClure, @hussainyk1.bsky.social, @polacklab.bsky.social we reveal how discretized representations in V1 predict suboptimal orientation discrimination. 🧪🧠🐭 This work reconciles neuro and psychometric curves
www.nature.com/articles/s41...

1 year ago 26 8 3 0
Post image

New paper in @brain1878.bsky.social: Healthy people under S-ketamine, an NMDAR antagonist, and people living with schizophrenia, a disorder associated with NMDAR hypofunction, spend more time in an external mode of perception - where noisy sensory signals override knowledge about the world.

1 year ago 26 8 1 2
Preview
The origin of color categories | PNAS To what extent does concept formation require language? Here, we exploit color to address this question and ask whether macaque monkeys have color ...

The origin of color categories | PNAS www.pnas.org/doi/10.1073/...

1 year ago 51 13 2 6

Check our latest in which we leverage shape metrics to compare neural geometry across regions, sessions or subjects and how their differences predict behavior.

w/ Nejatbakhsh, Duong, @sarah-harvey.bsky.social, Brincat, @siegellab.bsky.social, @earlkmiller.bsky.social & @itsneuronal.bsky.social

1 year ago 103 37 3 1
Post image

Paper shows very small LLMs can match or beat larger ones through 'deep thinking' - evaluating different solution paths - and other tricks. Their 7B model beats o1-preview on complex math by exploring 64 different solutions & picking the best one.

Test-time compute paradigm seems really fruitful.

1 year ago 157 20 3 4
Preview
Linking neural population formatting to function Animals capable of complex behaviors tend to have more distinct brain areas than simpler organisms, and artificial networks that perform many tasks tend to self-organize into modules (1-3). This sugge...

New results for a new year! “Linking neural population formatting to function” describes our modern take on an old question: how can we understand the contribution of a brain area to behavior?
www.biorxiv.org/content/10.1...
🧠👩🏻‍🔬🧪🧵
#neuroskyence
1/

1 year ago 232 82 2 7