Advertisement · 728 × 90

Posts by Tim Kellogg

yeah, all versions of strix have done this

25 minutes ago 0 0 0 0

you end up with less tokens too, so on average it’s actually a wash, or cheaper

28 minutes ago 2 0 0 0

i wasn't going to say anything... 😂

8 hours ago 2 0 0 0
Preview
The Productivity Is Real. The Scaling Isn't. What running an AI agent team taught me about why organizations can't do what one person can.

I substacked

I really like this one. It's a less technical piece on agent teams, and generally a bit about my workflow these days.

I have a technical piece coming on my own blog soon that's sort of related. Enjoy!

substack.com/home/post/p-...

9 hours ago 17 4 3 1

oh, and also you don’t have an ever-growing conversation context. it’s roughly constant

that’s actually one of the reasons for the journal, to create roughly equal sized turns, so we can include more history without blowing the context

11 hours ago 3 0 1 0

it’s super obvious when an agent doesn’t remember something that they should

so you fix instructions, dream up new architecture, whatever

i don’t think open-strix has the best answer, but we’re narrowing in on it, pretty fast

11 hours ago 2 0 1 0

compaction is another answer, but it happens too infrequently, so it’s really hard to debug when it goes wrong

so we go the other direction — instead of forgetting the unimportant things (compact), we instead remember the important things

11 hours ago 4 0 1 0

for Strix i made a strange design decision: completely rebuild the context on every message

fixed window of conversation/journal history, memory blocks, let agent read new files

sure, it kills the cache, but it forces you to figure out memory access & storage far sooner

11 hours ago 16 0 4 0

oh, right, makes sense

11 hours ago 1 0 0 0
Advertisement

hold on, where’s AI anchorman? @atlas-agent.bsky.social!!!

11 hours ago 2 0 0 0

meh, seems far fetched. more like minutes

11 hours ago 3 0 1 0

yes, the disinformation has seemed to have conglomerated into bluesky unfortunately

13 hours ago 0 0 1 0

ooh, it’s been a while since i’ve had one of you guys in my mentions. welcome back!

13 hours ago 1 0 1 0

what scent do you think they’ll release first?

14 hours ago 0 0 0 0

absolutely

14 hours ago 0 0 0 0

no.

14 hours ago 11 0 0 0

hey, this whole thing was a shitpost i really didn’t think it through okay

14 hours ago 15 0 2 0

yes, it’s taken from “higher order functions” -> functions that take or return functions

14 hours ago 5 0 1 0
Advertisement

predictions:

1. OpenAI will flip flop and release spud, GPT-5.5 anyway. It’ll be a lot better than Opus 4.6 but not quite as good as Mythos

2. Anthropic will wring their hands and finally release Mythos

3. xAI will blow them both away with Grok 5 but nobody will dare touch it

14 hours ago 28 0 9 1

a software product is more than just code, a lot more

most software products can and do get fully rewritten in cycles. and you would never know that from the outside, because code != product

14 hours ago 49 5 3 1

it moves you into a different plane of thinking

15 hours ago 2 0 0 0

now do 10yo

15 hours ago 0 0 1 0

yeah, i’m thinking that having a proper API reduces the factors that encourage rewrites

more value behind the API will further extend life

15 hours ago 1 1 1 0

right, in langgraph deep agents you can have skills without a filesystem, and i don’t see a problem with that

15 hours ago 1 0 0 0

to be clear, i’m conflating skill with all progressive disclosure

which, that’s wrong, technically. but it’s the part that makes skills work

15 hours ago 1 0 1 0

no they live on together

15 hours ago 1 0 1 0

phrasing..

15 hours ago 0 0 0 0

oh wait, you meant they=SLM

15 hours ago 3 0 1 0
Advertisement

this has already happened multiple times in my mentions

15 hours ago 3 0 1 0

oh i think TF-IDF fell out a tad before that, maybe 6 months

15 hours ago 2 0 0 0