No per-agent auto-worktree? This is the killer feature of Conductor, having to type `/worktree` into every new chat isn't really a resolution. Not even sure what selecting 'Worktree' for a new chat does
"having to type `/worktree` into every new chat isn't really a resolution"
I don't know what you're talking about. My experience with Cursor (before this new v3) is that new Cursor agent tabs / cloud agents already intelligently manage worktrees to prevent conflicts.
Wow, maybe something is wrong with my setup. In Cursor 3, I am clicking "New Agent" at the top left. My root repository is correctly listed on top of the composer, and I clicked the icon to the right of it and selected 'Worktree'. Then, I instruct the model to run `pwd` and tell me it's git status. It's always just on `main` in my root repository. I dug through the settings and couldn't find anything, and after finding this comment[0] on their forums gave up. Would you mind sharing a bit more about your setup/how it works?
Every company I've worked at has still had a few engineers who insist on working exclusively in the CLI with vim/emacs prior to AI. Every other engineer used some flavor of a desktop app ranging from more minimal editors to incredibly complex IDEs. I expect we land back on UIs long term.
On the above compared benchmarks is closer to other larger open weights models, and on par with GPT-OSS 120B, for which I also have a frame of reference.
Probably their auditors? Lying about this would be tantamount to (very serious) securities fraud. Not sure what you're basing on your allegations on besides "trust me bro"
Why would lying about having E2EE be securities (as in stock market) fraud? Would that make any lie ever told by a corporation equate to stock market fraud?
Most of the big AI/HPC clusters these systems are aimed at aren’t running regular PCIe Ethernet between nodes, they’re usually wired up with InfiniBand fabrics (HDR/NDR now, XDR soon)
The most egregious thing Perplexity did was to straight up ignore robots.txt. Cloudflare promise not to do that, so if we take their word for it, it's a quite different setup.
That said, I'm not fan of letting users forge whatever user agents they please. Instead, AIUI to opt-out of getting crawled I have to look for the existence of certain request headers[1].
reply