It looks like they did some worst case testing that was reassuring, so that it isn't Russian roulette? Any comments on that? I suppose their composite testing and temperature projections could also be wrong, and their trajectory changes might not be mitigating enough for the heat shield chunking, but that's a few different things all simultaneously being wrong for a catastrophic failure to occur.
The NASA engineers wanted to understand what would happen if large chunks of the heat shield were stripped away entirely from the composite base of Orion. So they subjected this base material to high energies for periods of 10 seconds up to 10 minutes, which is longer than the period of heating Artemis II will experience during reentry.
What they found is that, in the event of such a failure, the structure of Orion would remain solid, the crew would be safe within, and the vehicle could still land in a water-tight manner in the Pacific Ocean.
I think the point of the article is that there is no particular need to send humans in Artemis II.
Sure, they made tests. But it's not the same as trying in real conditions. The argument is that if they were able to predict everything with tests before the real flight, then Artemis I wouldn't have had those issues. But we know what happened.
Looks like they did some reassuring testing for the worst case scenario:
The Avcoat blocks, which are about 1.5 inches thick, are laminated onto a thick composite base of the Orion spacecraft. Inside this is a titanium framework that carries the load of the vehicle. The NASA engineers wanted to understand what would happen if large chunks of the heat shield were stripped away entirely from the composite base of Orion. So they subjected this base material to high energies for periods of 10 seconds up to 10 minutes, which is longer than the period of heating Artemis II will experience during reentry.
What they found is that, in the event of such a failure, the structure of Orion would remain solid, the crew would be safe within, and the vehicle could still land in a water-tight manner in the Pacific Ocean.
I wonder why he didnt just buy his way into this too. I guess Valve wasn't up for sale haha.
But he could have tried the VRChat folks, and Bigscreen. I guess he bought the Beat Saber folks, but he probably needed to buy a big game studio and maybe one with experience shipping successful MMORPGs.
I think we'll eventually move away from using these verbose documents, presentations, etc for communication. Just do your work, thinking, solving problems, etc while verbally dumping it all out into LLM sessions as you go. When someone needs to be updated on a particular task or project, there will be a way to give them granular access to those sessions as a sort of partial "brain dump" of yours. They can ask the LLM questions directly, get bullet points, whatever form they prefer the information in.
That way, thinking is communication! That's kind of why I loved math so much - it felt like I could solve a problem and succinctly communicate with the reader at the same time.
Of course it's going to be this government that goes and pokes the bubble that's propping up the economy, despite all the government's other shenanigans.
I believe the single core performance of the a18 pro is a 50% boost, but the multi core performance is about the same as the m1. I'm sure you're already taking the ram limitations into account for longevity.
While it's noisy and complicated for humans to read through, this session info is primarily for future AI to read and use as additional input for their tasks.
We could have LLMs ingest all these historical sessions, and use them as context for the current session. Basically treat the current session as an extension of a much, much longer previous session.
Plus, future models might be able to "understand" the limitations of current models, and use the historical session info to identity where the generated code could have deviated from user intention. That might be useful for generating code, or just more efficient analysis by focusing on possible "hotspots", etc.
Basically, it's high time we start capturing any and all human input for future models, especially open source model development, because I'm sure the companies already have a bunch of this kind of data.
That's exactly one of the reasons I've been archiving the sessions using DataClaw. The sessions can contain more useful information than the comments for humans.
TBH I don't think it's worth the context space to do this. I'm skeptical that this would have any meaningful benefits vs just investing in targeted docs, skills, etc.
I already keep a "benchmarks.md" file to track commits and benchmark results + what did/ did not work. I think that's far more concise and helpful than the massive context that was used to get there. And it's useful for a human to read, which I think is good. I prefer things remain maximally beneficial to both humans and AI - disconnects seem to be problematic.
Might not be worth it now, but might be in future. Not just for future LLMs, but future AI architectures.
I don't think the current transformers architecture is the final stop in the architectural breakthroughs we need for "AGI" that mimics human thought process. We've gone through RNN, LSTM, Mamba, Transformers, with an exponentially increasing amounts of data over the years. If we want to use similar "copy human sequences" approaches all the way to AGI, we need to continuously record human thoughts, so to speak (and yes, that makes me really queasy).
So, persisting the session, that's already available in a convenient form for AI, is also about capturing the human reasoning process during the session, and the sometimes inherent heuristics therein. I agree that it's not really useful for humans to read.
I just don't really see the point in hedging like that tbh. I think you could justify almost anything on "it could be useful", but why pay the cost now? Eh.
> While it's noisy and complicated for humans to read through, this session info is primarily for future AI to read and use as additional input for their tasks.
Context rot is very much a thing. May still be for future agents. Dumping tens/hundreds of thousand of trash tokens into context very much worsen the performance of the agent
It's just noise for AI too. There is no reason to be lazy with context management when you can simply ask the AI to write the summary of the session. But even that is hardly useful when AI can just read the source of truth which is the code and committed docs
Q: Will you turn off the tool if they violate the rules?
@sama: Yes, we will turn it off in that very unlikely event, but *we believe the U.S. government is an institution that does its best to follow law and policy.*
What we won't do is turn it off because we disagree with a particular (legal military) decision. We trust their authority.
How can anyone trust a guy who says this after the past year?!
I bet Sam secretly pledged to DoD that the red lines were only temporary, for optics and to calm employees at the all hands meeting.
A few months down the line, OpenAI will quietly decide that their next model is safe enough for autonomous weapons, and remove their safeguard layer. The mass surveillance enablement might be an indirect deal through Palantir.
The fact they were able to strike a deal in the first place hours after Anthropic was declared a supply chain risk should make this obvious. Their employees are smart people, the only way they can reason past this is their compensation.
The NASA engineers wanted to understand what would happen if large chunks of the heat shield were stripped away entirely from the composite base of Orion. So they subjected this base material to high energies for periods of 10 seconds up to 10 minutes, which is longer than the period of heating Artemis II will experience during reentry.
What they found is that, in the event of such a failure, the structure of Orion would remain solid, the crew would be safe within, and the vehicle could still land in a water-tight manner in the Pacific Ocean.
reply