tal

joined 2 years ago
[–] tal@lemmy.today 5 points 3 months ago

Third it has network effect going for it. Nobody is going to watch videos on your platform if there’s only a couple dozen of them total. The sheer size and scope of YouTube means no matter what you’re looking for you can find something to watch.

Yeah, though I think that you could avoid some of that with a good cross-video-hosting service search engine, as I don't think that most people are engaging in the social media aspect of YouTube. YouTube doesn't have a monopoly on indexing YouTube videos.

But the scale doesn't hurt them, that's for sure.

[–] tal@lemmy.today 3 points 3 months ago

I wasn't a regular follower in recent years, so I'm reaching a bit further back, but yeah, I recall a steady flow of people submitting general questions and mods removing them. I'd have probably just treated it like a desire path (!desire_paths@sh.itjust.works, BTW)


if that's how people want to walk, maybe just a sign that it's easier to just build a path there.

thinks

I suppose that there were some changes that could have happened in the move from Reddit.

There was also a collection of people who didn't want to copy the "*porn" convention from Reddit for attractive-but-non-pornographic pictures of things (that one doesn't bother me, but I do understand people who are uncomfortable about it and wanted to shelve it in the move). Like, their workplace may not care about people looking at landscape pictures, but gets twitchy about anything remotely porn-related.

There are also some pretty obscure jokes that came from long-ago Reddit drama or jokes that probably make the Threadiverse more-complicated to navigate for people who weren't in on the joke from years back. Like the "inversion" communities, like trees/MarijuanaEnthusiasists (!trees@lemmy.world and !MarijuanaEnthusiasts@lemmy.world) or worldnews/anime_titties (!worldnews@lemmy.world and !anime_titties@lemmy.world, though it looks like eventually, worldnews went back to being actual world news both on Reddit and here). Or /r/superbowl (!superbowl@lemmy.world), though I think that that one, at least, someone can figure out if they stumble into it. Might have been a good argument that we should have adopted more-conventional naming. But I think that the bigger concern in the big move was getting things up-and-running, rather than trying to rearchitect everything.

[–] tal@lemmy.today 22 points 3 months ago* (last edited 3 months ago) (2 children)

IMHO, the real problem is that the community is poorly-named. It should be "ThoughtfulDiscussion" or something. The name suggests a general forum to ask any question. And so, well, people do.

The /r/askreddit subreddit had the same problem as !asklemmy@lemmy.world does, as I recall.

EDIT: I'd add that I think that there's actually a better argument for a general "ask questions" community on the Threadiverse than on Reddit, at least as things stand in 2025, because the userbase is smaller, so it's hard to get many people in a lot of the niche forums. Like, sure, if you want to ask a question about Linux or about a video game, there are more-appropriate communities. But...suppose you want to ask a question about, say, fly-fishing? I haven't looked, but I'll bet that there isn't even a fly-fishing community out there yet.

EDIT2: !casualconversation@piefed.social is sorta-kinda for general posts that are intended to spark conversations, and the content there might be somewhat-closer to what you're looking for, if you want content that people would actually talk about. I don't know if I'd call all of that "thought-provoking", but I think that stuff there is better at starting back-and-forth conversations, rather than just getting a one-off answer.

[–] tal@lemmy.today 2 points 3 months ago

I did see some depth=1 or something like that to get only a certain depth of git commits but thats about it.

Yeah, that's a shallow clone. That reduces what it pulls down, and I did try that (you most-likely want a bit more, probably to also ask to only pull down data from a single branch) but back when I was crashing into it, that wasn't enough for the Cataclysm repo.

It looks like it's fixed as of early this year; I updated my comment above.

[–] tal@lemmy.today 1 points 3 months ago* (last edited 3 months ago) (2 children)

Thanks. Yeah, I'm pretty sure that that was what I was hitting. Hmm. Okay, that's actually good


so it's not a git bug, then, but something problematic in GitHub's infrastructure.

EDIT: On that bug, they say that they fixed it a couple months ago:

This seems to have been fixed at some point during the last days leading up to today (2025-03-21), thanks in part to @MarinoJurisic 's tireless efforts to convince Github support to revisit this problem!!! 🎉

So hopefully it's dead even specifically for GitHub. Excellent. Man, that was obnoxious.

[–] tal@lemmy.today 3 points 3 months ago* (last edited 3 months ago) (4 children)

A bit of banging away later


I haven't touched Linux traffic shaping in some years


I've got a quick-and-dirty script to set a machine up to temporarily simulate a slow inbound interface for testing.

slow.sh test script

# !/bin/bash
# Linux traffic-shaping occurs on the outbound traffic.  This script
# sets up a virtual interface and places inbound traffic on that virtual
# interface so that it may be rate-limited to simulate a network with a slow inbound connection.
# Removes induced slow-down prior to exiting.  Needs to run as root.

# Physical interface to slow; set as appropriate
oif="wlp2s0"

modprobe ifb numifbs=1
ip link set dev ifb0 up
tc qdisc add dev $oif handle ffff: ingress
tc filter add dev $oif parent ffff: protocol ip u32 match u32 0 0 action mirred egress redirect dev ifb0

tc qdisc add dev ifb0 root handle 1: htb default 10
tc class add dev ifb0 parent 1: classid 1:1 htb rate 1mbit
tc class add dev ifb0 parent 1:1 classid 1:10 htb rate 1mbit

echo "Rate-limiting active.  Hit Control-D to exit."
cat

# shut down rate-limiting
tc qdisc delete dev $oif ingress
tc qdisc delete dev ifb0 root
ip link  set dev ifb0 down
rmmod ifb

I'm going to see whether I can still reproduce that git failure for Cataclysm on git 2.47.2, which is what's in Debian trixie. As I recall, it got a fair bit of the way into the download before bailing out. Including the script here, since I think that the article makes a good point that there probably should be more slow-network testing, and maybe someone else wants to test something themselves on a slow network.

Probably be better to have something a little fancier to only slow traffic for one particular application


maybe create a "slow Podman container" and match on traffic going to that?


but this is good enough for a quick-and-dirty test.

[–] tal@lemmy.today 11 points 3 months ago* (last edited 3 months ago) (8 children)

This low bandwidth scenario led to highly aggravating scenarios, such as when a web app would time out on [Paul] while downloading a 20 MB JavaScript file, simply because things were going too slow.

Two major applications I've used that don't deal well with slow cell links:

  • Lemmyverse.net runs an index of all Threadiverse instances and all communities on all instances, and presently is an irreplaceable resource for a user on here who wants to search for a given community. It loads an enormous amount of data for the communities page, and has some sort of short timeout. Whatever it's pulling down internally

I didn't look


either isn't cached or is a single file, so reloading the page restarts from the start. The net result is that it won't work over a slow connection.

  • This may have been fixed, but git had a serious period of time where it would smash into timeouts and not work on slow links, at least to github. This made it impossible to clone larger repositories; I remember failing trying to clone the Cataclysm: Dark Days Ahead repository, where one couldn't even manage a shallow clone. This was greatly-exacerbated by the fact that git does not presently have the ability to resume downloads if a download is interrupted. I've generally wound up working around this by git cloning to a machine on a fast connection, then using rsync to pull a repository over to the machine on a slow link, which, frankly, is a little embarrassing when one considers that git really is the premier distributed VCS tool out there in 2025, and really shouldn't need to rely on that sort of workaround.
[–] tal@lemmy.today 27 points 3 months ago (10 children)

I'm confident that we could set up permanent human habitation on the Moon or on Mars with our current level of technology, and that's featured pretty prominently in sci-fi.

I don't know if it would actually provide a cost-effective return, but I do think that it'd be interesting to see happen in my lifetime.

[–] tal@lemmy.today 2 points 3 months ago

Ah, gotcha. You'll still have an option in that case


if you go to Amazon (or Monoprice...traditionally, they were my go-to spot for cables, but I haven't tried pricing them against Amazon recently), they'll also have male-to-female extension cables for USB and HDMI. I keep a USB extension cable in the car, as I normally want short cables, but every now and then, I want to put something further away.

[–] tal@lemmy.today 8 points 3 months ago* (last edited 3 months ago) (2 children)

sabotage

Microsoft's interest in Nokia was being able to compete with what is now a duopoly between Google and Apple in phones. They wanted to own a mobile platform. I am very confident that they did not want their project to flop. That being said, they'll have had their own concerns and interests. Maybe Nokia would have done better to go down the Apple or Google path, but for Microsoft, the whole point was to get Microsoft-platform hardware out there.

[–] tal@lemmy.today 3 points 3 months ago* (last edited 3 months ago) (2 children)

(Side note, a longish cable is appreciated, but not required, I think the current length is around 5 ft, which is just about enough)

Almost everything is going to be USB or HDMI, and will virtually always have a replaceable cable, so you can get whatever length you want, within the spec of that protocol.

EDIT: If you really want a long cable, something that exceeds what copper can do, you can even get cables that will contain an optical transceiver and contain a fiber optic strand


I have a long USB cable like this. I assume that you don't need that, though, if you have the computer and the webcam in more-or-less the same place.

[–] tal@lemmy.today 20 points 3 months ago (6 children)

And Amazon says it will help train 4 million people in AI skills and “enable AI curricula” for 10,000 educators in the US by 2028, while offering $30 million in AWS credits for organizations using cloud and AI tech in education.

So, at some point, we do have to move on policy, but frankly, I have a really hard time trying to predict what skillset will be particularly relevant to AI in ten years. I have a hard time knowing exactly what the state of AI itself will be in ten years.

Like, sure, in 2025, it's useful to learn the quirks and characteristics of LLMs or diffusion models to do things with them. I could sit down and tell people some of the things that I've run into. But...that knowledge also becomes obsolete very quickly. A lot of the issues and useful knowledge for, working with, say, Stable Diffusion 1.5 are essentially irrelevant as regards Flux. For LLMs, I strongly suspect that there are going to be dramatic changes surrounding reasoning, and retaining context. Like, if you put education time into training people on that, you run the risk that they don't learn stuff that's relevant over the longer haul.

There have been major changes in how all of this works over the past few years, and I think that it is very likely that there will be continuing major changes.

view more: ‹ prev next ›