yes_this_time

joined 2 years ago
[–] yes_this_time@lemmy.world 8 points 6 days ago (1 children)

Yeah, after reading a bit into it. It seems like most of the work is up front, pre filtering and classifying before it hits the model, to your point the model training part is expensive...

I think broadly though, the idea that they are just including the kitchen sink into the models without any consideration of source quality isn't true

[–] yes_this_time@lemmy.world 4 points 6 days ago

Good points. What's novel information vs. wrong information? (And subtly wrong is harder to understand than very wrong)

At some point it's hitting a user who is giving feedback, but I imagine data lineage once it gets to the end user its tricky to understand.

[–] yes_this_time@lemmy.world 13 points 6 days ago (8 children)

If I'm creating a corpus for an LLM to consume, I feel like I would probably create some data source quality score and drop anything that makes my model worse.

[–] yes_this_time@lemmy.world 3 points 1 week ago

The bans are for under 16s, not just 7 year olds. Parents don't control all internet activity for 15 years, at that age they are going to have some autonomy outside of the house.

I'm not sure there is a direct irl analog when it comes to controlling digital spaces, since they are personal by nature. and I think this is where the debate comes in.

Should parents be following their teenage child into every store to make sure they aren't buying alcohol?

I get the concern with providing social media companies a government ID, I certainly would never give them one! I would just not use them. But they provide net negative value in my opinion so no loss.

I like the idea of FOSS parental controls.

[–] yes_this_time@lemmy.world 2 points 1 week ago* (last edited 1 week ago)

Not to say it's never the parents fault...

[–] yes_this_time@lemmy.world 8 points 1 week ago (12 children)

There is precedence though. We age gate: nicotine, alcohol, gambling etc..

we shouldnt expect parents to be monitoring children 24/7. actually, as children get older they should be given freedoms, parents have the right to expect our society has some guardrails.

[–] yes_this_time@lemmy.world 14 points 2 weeks ago* (last edited 2 weeks ago)

Eh, don't be rude. You are likely thinking of single gene mutations or other clear well defined problems.

My mind was more on polygenic diseases or genes with variable expressiveness. Where humans being humans we target things where we don't completely understand the outcomes.

We screen for chromosomal abnormalities I don't have a problem with that for example.

[–] yes_this_time@lemmy.world 22 points 2 weeks ago (17 children)

There is a strong possibility we would also get it wrong. Diversity is a strength. Who knows what tomorrow brings.

[–] yes_this_time@lemmy.world 3 points 2 months ago

It makes me a bit sad that there is a whole article on a (very likely) mirage

[–] yes_this_time@lemmy.world 4 points 3 months ago (7 children)

500 million was specific to Claude Code, they are at 5 billion annual run rate and growing

[–] yes_this_time@lemmy.world 0 points 4 months ago (1 children)

I mean, I agree that a lot of money was spent training some of these models - and I personally wouldn't invest in an ai based company. The economics dont make sense.

However, worst case, self hosted open source models have got pretty good, and I find it unlikely that progress will simply stop. Diminishing returns from scaling data yes, but there will still be optimizations all through the pipeline.

That is to say, LLMs will continue to have utility regardless if Open AI and Anthropic are around long term.

[–] yes_this_time@lemmy.world 3 points 4 months ago (3 children)

Some people are finding value in LLMs, that doesn't mean LLMs are great at everything.

Some people have work to do, and this is a tool that helps them do their work.

view more: next ›