Hacker Newsnew | past | comments | ask | show | jobs | submit | danielmarkbruce's commentslogin

We already have a mechanism. Contract law.

Just stop thinking of software products. There are a million businesses offering solid value propositions that need a lot of software to run, but software isn't the product.

Path dependence is a thing.

Maybe, but how exactly are you defining "code quality" ?

This is an absurd statement. There are many complex undertakings in sport where even the very best get better with practice and can't tell you why. In fact, the ones who think they can tell you why are the one's to be most skeptical of.

You are just making stuff up or regurgitating material from a pop science book.


They can't tell you (not everyone is eloquent), but they sure know why. Struggling to put something in word is not the same as not knowing.

Much of human behavior is evolved so that we don't understand why. For example human morality is an evolved trait, but you wouldn't know it.

Please explain walking to me so that I can explain it to a person who forgot how to walk such that he can walk after the explanation.


Nope, they don't.

"Designed" is a bit strong. We "literally" couldn't design programs to do the interesting things LLMs can do. So we gave a giant for loop a bunch of data and a bunch of parameterized math functions and just kept updating the parameters until we got something we liked.... even on the architecture (ie, what math functions) people are just trying stuff and seeing if it works.


> We "literally" couldn't design programs to do the interesting things LLMs can do.

That's a bit of an overstatement.

The entire field of ML is aimed at problems where deterministic code would work just fine, but the amount of cases it would need to cover is too large to be practical (note, this has nothing to do with the impossibility of its design) AND there's a sufficient corpus of data that allows plausible enough models to be trained. So we accept the occasionally questionable precision of ML models over the huge time and money costs of engineering these kinds of systems the traditional way. LLMs are no different.


Saying ML is a field where deterministic code would work just fine conveniently leaves out the difficult part - writing the actual code.... Which we haven't been able to do for most of the tasks at hand.

What you are saying is fantasy nonsense.


They did not leave it out.

> but the amount of cases it would need to cover is too large to be practical (note, this has nothing to do with the impossibility of its design)


It's not only too large - we can't even enumerate all the edge cases, let alone handle them. It's too difficult.

Using your logic, we don’t need quantum computers to break encryption, we could just use pen and paper.

And all you have to do is write an infinite amount of code to cover all possible permutations of reality! No big deal, really.

> would work just fine, but the amount of cases it would need to cover is too large to be practical

So it doesn't work.


It is impossible to design even in a theoretical sense if functional requirements consider matters such as performance and energy consumption. If you have to write petabytes of code you also have to store and execute it.

[flagged]


I'm a psychiatry resident who has been into ML since... at least 2017. I even contemplated leaving medicine for it in 2022 and studied for that, before realizing that I'd never become employable (because I could already tell the models were getting faster than I am).

You would be sorely mistaken to think I'm utterly uninformed about LLM-research, even if I would never dare to claim to be a domain expert.


No, it's not common for the startup itself to make capital calls. The phrase (and your link) refers to capital calls made by VC firms to their limited partners. Same thing in PE.


You don't have an email address listed - could you ping me at my username at googles popular email service to chat on tools for institutional investing?


vicchenai is an LLM bot.


UI + functions on top of schema if you've designed the schema well. Otherwise, it's a whole other thing.


This very much depends on where you work... and basically isn't true for most people. It's extremely true for some people.


Rule 3 is still very much real. Fancy fast algorithms often have other trade-offs. The best algorithm for the job is the one that meets all requirements well... Big-O is one aspect, data is another, determinism of the underlying things that are needed (dynamic memory allocation, etc) can be another.

It is important to remember that the art of sw engineering (like all engineering) lives in a balance between all these different requirements; not just in OPTIMIZE BIG-O.


Sure but the default (and usually correct) assumption when working at google (as an example) is basically "all numbers are big", so you you have to cluey about algorithms and data structures and not default to brute forcing something.

At 99% of shops it should be the other way around .


Even when you are working with large numbers, most numbers are usually small. Most of the code is probably not dealing with the large things, and a large thing may consist of a large number of instances that are individually small.

I've personally found it useful to always have concrete numbers in mind. An algorithm or data structure designed for N will probably be fine for N/10 and 10N, but it will often be inefficient for N/1000 or 1000N.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: