Hacker Newsnew | past | comments | ask | show | jobs | submit | monooso's commentslogin

For the avoidance of doubt, you won't donate to Thunderbird because you disagree with the following policy?

> We do not discriminate on the basis of race [...], religion [...], gender, gender identity, gender expression, color, national origin, pregnancy, ancestry, domestic partner status, disability, sexual orientation, age, genetic predisposition, medical condition, marital status, citizenship status, military or veteran status, or any other basis covered by applicable laws. Mozilla will not tolerate discrimination or harassment based on any of these characteristics or any other unlawful behavior, conduct, or purpose.


Yes, that is correct. Discrimination is already illegal in hiring. Spelling it out so absurdly verbosely is just virtue signalling. If you're a remote developer, nobody cares about your colour or sex. Except at Mozilla, where people have their pronouns in Bugzilla.

To be clear, I fully support the right to be and feel and think whatever you want, but don't expect me to care about it, and this endless signposting of identity is tiresome.


You say you don't care about these identities, yet you’re willing to let a tool you (presumably) value lose funding over a text block you find 'verbose.' It seems you care enough about the signposting to let it outweigh the software's actual utility.

get over yourself

> Anthropic is up to $30B annual recurring revenue. I wish I had failing business models like that.

And profit? A company can have $300B annual revenue, and still be a failing business if it's making a loss.

Somewhere along the line we seem to have forgotten this basic fact. Eventually there will be no more rounds of funding to feed the fire.


Anthropic has raised $64B in total since they were founded.

Even if you say we are going to measure profit in the very special hacker news way of looking at money taken in from customer revenue against money invested and we say they can't do things like counting building data centers or buying GPUs as capital expenses and instead have to count them against profit then in 2 years time they will have made more money than they have taken in investment.

That is extraordinary.


The numbers are indeed extraordinary. But I still don't agree with this statement.

> ...in 2 years time they will have made more money than they have taken in investment.

In two years' time they will have generated more revenue, that is not the same as "making money."


Costs can always be optimized, revenue is much harder to optimize.

I haven't seen anyone claim that applies to infrastructure or compute.

Since apparently LLMs have also conquered physics, “Claude, transmute this lead to gold for me.”

Yeah, it's almost like the point I was making is that everyone is overselling AI agents' capabilities.

I’m sure someone is out there claiming that AI is going to solve all your business’s problems no matter what they are. Remotely sane people are saying it will solve (or drastically improve) certain classes of problems. 3x code? Sure. 3x the physical hardware in a data center? Surely not.

Implying that software is somehow divorce from Infrastructure/compute efficiency and utilization isn't a claim I've seen many make either.

I haven't used the app yet, but from the screenshots it appears to follow the GNOME Human Interface Guidelines [1].

You may disagree with said guidelines, of course, but the author can't reasonably be criticised for following the platform standards.

[1]: https://developer.gnome.org/hig/


No idea why you're being downvoted. I've done my bit to redress the balance, I hope others do the same.

You have no evidence that it was.

That's just writing. I frequently write like that.

This insistence that certain stylistics patterns are "tell-tale" signs that an article was written by AI makes no sense, particularly when you consider that whatever stylistic ticks an LLM may possess are a result of it being trained on human writing.


These are just some of the good examples I found.

My hunch that this is substantially LLM-generated is based on more than that.

In my head it's like a Bayesian classifier, you look at all the sentences and judge whether each is more or less likely to be LLM vs human generated. Then you add prior information like that the author did the research using Claude - which increases the likelihood that they also use Claude for writing.

Maybe your detector just isn't so sensitive (yet) or maybe I'm wrong but I have pretty high confidence at least 10% of sentences were LLM-generated.

Yes, the stylistic patterns exist in human speech but RLHF has increased their frequency. Also, LLM writing has a certain monotonicity that human writing often lacks. Which is not surprising: the machine generates more or less the most likely text in an algorithmic manner. Humans don't. They wrote a few sentences, then get a coffee, sleep, write a few more. That creates more variety than an LLM can.

Fun exercise: https://en.wikipedia.org/wiki/Wikipedia:AI_or_not_quiz


Here's an alternative way of thinking about this...

Someone probably expended a lot of time and effort planning, thinking about, and writing an interesting article, and then you stroll by and casually accuse them of being a bone idle cheat, with no supporting evidence other than your "sensitive detector" and a bunch of hand-wavy nonsense that adds up to naught.


To start, this is more or less an advertising piece for their product. It's pretty clear that they want to sell you Allium. And that's fine! They are allowed! But even if that was written by a human, they were compensated for it. They didn't expend lots of effort and thinking, it's their job.

More importantly, it's an article about using Claude from a company about using Claude. I think on the balance it's very likely that they would use Claude to write their technical blog posts.


> They didn't expend lots of effort and thinking, it's their job.

Your job doesn't require you to think or expend effort?


While I agree with the sentiment, using AI to write the final draft of the article isn’t cheating. People may not like it, but it’s more a stylistic preference.

Using AI and a human byline is 100% cheating.

Yeah I agree. Don't tell me you authored something when claude did the majority of the writing. Use claude if you want, but don't pretend you wrote the content when you didn't.

I also hate this style of plastic, pre-digested prose. Its soulless and uninteresting. Maybe I've just read too much AI slop. I associate this writing style with low quality, uninteresting junk.


Yet another way the mere possibility of AI/LLM being involved diminishes the value of ALL text.

If there is constant vigilance on the part of the reader as to how it was created, meaning and value become secondary, a sure path to the death of reading as a joy.


Those aren’t good examples - that’s just LLMs living for free in your head.

I am reminded of the Simpsons episode in which Principal Skinner tries to pass off the hamburgers from a near-by fast food restaurant for an old family recipe, 'steamed hams,' and his guest's probing into the kitchen mishaps is met with increasingly incredible explanations.

See also: “I'm Kenyan. I Don't Write Like ChatGPT. ChatGPT Writes Like Me” by Marcus Olang', https://marcusolang.substack.com/p/im-kenyan-i-dont-write-li...

For what it’s worth, Pangram reports that Marcus’ article is 100% LLM-written: https://www.pangram.com/history/640288b9-e16b-4f76-a730-8000...


In theory, wouldn't be too hard be to settle the question if whether he used ChatGPT to write it: get Olang to write a few paragraphs by hand, then have people judge (blindly) if it's the same style as the article. Which one sounds more like ChatGPT.

The times I've written articles, and those have gone through multiple rounds of reviews (by humans) with countless edits each time, before it ends up being published, I wonder if I'd pass that test in those cases. Initial drafts with my scattered thoughts usually are very different from the published end results, even without involving multiple reviewers and editors.

When people judge blindly, the are more likely to think the human is the AI and the AI is the human.

73% judged GPT 4.5 (edit: had incorrectly said 4o before)to be the human.

https://arxiv.org/abs/2503.23674

Not only are people bad at judging this, but are directionally wrong.


There is research showing the contrary that is far more convincing:

> Our experiments show that annotators who frequently use LLMs for writing tasks excel at detecting AI-generated text, even without any specialized training or feedback. In fact, the majority vote among five such “expert” annotators misclassifies only 1 of 300 articles, significantly outperforming most commercial and open-source detectors we evaluated even in the presence of evasion tactics like paraphrasing and humanization.

https://arxiv.org/html/2501.15654v2


Great find, I've submitted this preprint as a standalone item: https://news.ycombinator.com/item?id=47678270

I’m so glad the witch hunt has moved on to phrasing so I get less grief for my em dashes.

I hate that I can’t write em dashes freely anymore without people accusing the writing of being AI generated.

Even though they are perfect for usage in writing down thoughts and notes.


One thing you can try⸺admittedly it's not quite correct⸺is replacing them with a two-em dash. I've never seen an AI use one, and it looks pretty funky.

Since the advantage of standards is that there are so many to choose from, one lesser-used but still regionally acceptable approach (e.g. https://www.alberta.ca/web-writing-style-guide-punctuation#j...) is to use en-dashes offset with spaces.

I have nothing against em dashes. As long as your writing is human, experienced readers will be able to tell it's human. Only less experienced ones will use all or nothing rules. Em dashes just increase the likelihood that the text was LLM generated. They aren't proof.

That nuance is lost on the majority of anti-AI folks who’ve learned they get positive social reactions by declaring essentially everything to be AI written and condemnable.

“An em dash… they’re a witch!”… “it’s not just X, it’s Y… they’re a witch!”


> anti-AI folks who’ve learned they get positive social reactions by declaring essentially everything to be AI written and condemnable.

that's a strawman alright; all the comments complaining how they can't use their writing style without being ganged up on are positive karma from my angle, so I'm not sure the "positive social reactions" are really aligned with your imagination. Or does it only count when it aligns with your persecution complex?


You have the same problem apparently. You think it’s okay to go witch hunting and accuse people with no real evidence.

Evidently there are no experienced readers who post AI accusations.

Same weight as "there are no experienced men who'll ask a woman if she's pregnant."

Why do you care what others accuse you of?

No, it’s pretty obviously AI written. Not sure why you’re running so much interference for them…are you affiliated with this company?

> This is the dumbest, most pointless military conflict in American history.

You've picked a high bar there.


Genuine question: is Fox News a credible, unbiased source for these figures in particular?

Their poll is, and this is not the only poll I picked. CNN has multiple polls listed on its site, and this is the first I found which gives the party breakdown.

I believe the phrase you're reaching for is "carrying out extra-judicial executions of civilians alleged to have been smuggling drugs."

Or shorter.. NAZIS!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: