Hacker Newsnew | past | comments | ask | show | jobs | submit | tontinton's commentslogin

or you can use a non vibe designed efficient Rust TUI coding agent made by yours truly, all my coworkers use it too :) called https://maki.sh!

lua plugins WIP


Thanks!

Currently the subagent chat windows don't allow to inject user messages like the main window, I want to change that soon though.

Regarding tiered models, it currently caps the model use to the current tier you're on, so no it can't upgrade from haiku to opus suddenly. The reasoning for that is that if you selected haiku, you probably don't want to pay for opus by accident.


Is this a good time to tell you guys that there are many more alternatives, like my very own https://maki.sh?

Shameless plug, yes, but it's free, so I think it's fair :)


This looks cool. Does it support working with local models?


Yes, you can use the dynamic provider to work with any local model you want (see scripts/providers)


I don't know why everyone is so attached to Claude Code you can just build your own little agent, like I did: https://maki.sh/

It will 100% be better than the 500k lines of code junk that is CC.


Something I've been working on out of frustrations with the existing AI coder CLIs


Yeah we all converge to the same workflow, in my ai coding agent I'm working on now, I've added an "index" tool that uses tree-sitter to compress and show the AI a skeleton of a code file.

Here's the implementation for the interested: https://github.com/tontinton/maki/blob/main/maki-code-index%...


I'm curious, what does your workflow look like? I saw a plan prompt there, but no specs. When you want to change something, implement a new feature etc, do you just prompt requirements, have it write the plan and then have it work on it?


Oh, that's great.

I've always wanted to explore how to fit tree-sitter into this workflow. It's great to know that this works well too.

Thanks for sharing the code.

(Here is the AIPack runtime I built, MIT: https://github.com/aipack-ai/aipack), and here is the code for pro@coder (https://github.com/aipack-ai/packs-pro/tree/main/pro/coder) (AIPack is in Rust, and AI Packs are in md / lua)


Is it similar to rtk? Where the output of tool calls is compressed? Or does it actively compress your history once in a while?

If it's the latter, then users will pay for the entire history of tokens since the change uncached: https://platform.claude.com/docs/en/build-with-claude/prompt...

How is this better?


This is a bit more akin to distill - https://github.com/samuelfaj/distill

Advantage of SML in between some outputs cannot be compressed without losing context, so a small model does that job. It works but most of these solutions still have some tradeoff in real world applications.


We do both:

We compress tool outputs at each step, so the cache isn't broken during the run. Once we hit the 85% context-window limit, we preemptively trigger a summarization step and load that when the context-window fills up.


> we preemptively trigger a summarization step and load that when the context-window fills up.

How does this differ from auto compact? Also, how do you prove that yours is better than using auto compact?


For auto-compact, we do essentially the same Anthropic does, but at 85% filled context window. Then, when the window is 100% filled, we pull this precompaction + append accumulated 15%. This allows to run compaction instantly


Very cool, have you taken a look into what TigerBeetle does with VSR (and why they chose it instead of raft)?


Yes I’ve read through TigerBeetle’s VSR design and their rationale for not using Raft.

VSR makes a lot of sense for their problem space: fixed schema, deterministic state machine, and a very tight control over replication + execution order.

Ayder has a different set of constraints: - append-only logs with streaming semantics - dynamic topics / partitions - external clients producing arbitrary payloads over HTTP

Raft here is a pragmatic choice: it’s well understood, easier to reason about for operators, and fits the “easy to try, easy to operate” goal of the system.

That said, I think VSR is a great example of what’s possible when you fully own the problem and can specialize aggressively. Definitely a project I’ve learned from.


Can I select multiple receivers concurrently, similar to a select in Linux?


Learn about the data structures & algorithms that make up modern log search engines like Elasticsearch.


Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: