• 0 Posts
  • 63 Comments
Joined 1 year ago
cake
Cake day: August 2nd, 2023

help-circle





  • We’re likely to see a variant of Moore’s law when it comes to satellites. Launch costs will keep going down. Right now we have Starlink with a working satellite internet system and China with a nascent one. As the costs come down we’ll likely see more and more countries, companies, organizations and individuals will be able to deploy their own systems.

    A government would need to negotiate with every provider to get them to block signals over their country. Jamming is always hard. You could theoretically jam all communications or communications on certain frequency bands but it’s not clear how you would selectively jam satellite internet.


  • There’s a much bigger story here.
    Think about how hard it was to discover this access point. Even after it was reported and there was a known wi-fi network and the access point was known to be on a single ship, it took the Navy months to find it.

    Starlink devices are cheap and it will be nearly impossible to detect them at scale. That means that anyone can get around censors. If the user turns off wi-fi, they’ll be nearly impossible to detect. If they leave wi-fi on in an area with a lot of wi-fi networks it will also be nearly impossible to detect. A random farmer could have Starlink in their hut. A dissident (of any nation) could hide the dish behind their toilet.

    As competing networks are launched, users will be able to choose from the least restricted network for any given topic.


  • The effect is mostly from the total number of computer users increasing.

    That is, the total number of “tech-savvy” users keeps increasing (https://datausa.io/profile/cip/computer-science-110701) but the number of “non-tech-savvy” computer users has absolutely exploded (https://semiconalpha.substack.com/p/global-semiconductor-sales-increase) (that actually undercounts computers since every dollar in 2020 buys you much more computer power than a dollar in 1987)

    You had to pass a nerd gauntlet just to get online in the 80’s or 90’s that meant that everyone you met online had also passed that gauntlet and was tech savvy. Even if you looked in the social usenet groups, a lot of non-technical users were just filtered out. So it looked like everyone was tech savvy but that’s because we were sampling a tiny, tech-savvy portion of the population.

    Now anyone can get online. The tech savvy gen-zers are still there but their hidden in a sea of non-technical users. If you go to places like Github or Hackernews (or even more specifically technical fora), you’ll find plenty of enthusiastic young people poking at technology and trying to make it better. They no longer have to mess around with autoexec.bat and config.sys to get their mouse working but they can (and do) get a bunch of Jupyter notebooks and start playing around with Tensorflow.

    A great modern example of this is 3-D printing. Modern 3-D printers suck. If you’re a big company you can get super expensive 3-D printers that take up giant rooms and need a team of experts to run. If you’re a home user you can get a cheap FDM printer but you best be prepared to tinker with it. The first thing most people do with their Ender is print mods for their Ender. Bambu Labs is a big improvement but they also attract a lot of users who at least could mod their printer https://forum.bambulab.com/c/bambu-lab-x1-series/user-mods/19

    Some day we may have little boxes like in “Diamond Age”. Kids in the future may not even know about crap like bed adhesion and stringing and they’ll concentrate on whatever the new problems are revealed once the current ones are taken care of.



  • I think that true “tech-savvyness” isn’t really a generational thing.

    Some people are just really curious about how stuff works. When they see something they aren’t satisfied with, “Just do it.” or “Shit just works.” They want to know how and why it works. When you hand those people a computer, machine or flower they’ll poke at it and try to understand it better.

    It’s not clear that typing skills are actually needed for that.

    I max out at around 80-100 WPM but I only sustain that when I’m transcribing something. When I need to learn about technology, it’s much more about reading than typing. When I actually need to do some coding, I spend much more time staring at the screen and looking up stuff on Stackoverlow than I do actually typing.

    Most of Z is not savvy at all, just like with every generation. And just like with every generation, some of them will push the envelope of technology. I doubt that lack of typing will slow those folks down.






  • It’s hard to draw meaningful conclusions form a single 4 year period. There have been several instances of corruption (and significant externalized costs) in private firms that went on for much longer than 4 years.

    I agree that there is a lot of corruption in government but there’s a long gap between that and no accountability. We see various forms of government accountability on a regular basis; politicians lose elections, they get recalled, and they sometimes even get incarcerated. We also have multiple systems designed to allow any citizen to influence government.

    None of these systems and safeguards are anywhere close to perfect but it must be better than organizations that don’t even have these systems in the first place.


  • What makes governments any more susceptible to corruption than a private organization?

    I’m not actually talking about governments having absolute control. That’s a pretty extreme scenario to jump to from from the question of if it’s better for a private company or a government to control search.

    Right now we think Google is misusing that data. We can’t even get information on it without a leak. The government has a flawed FOIA system but Google has nothing of the sort. The only way we’re protected from corruption at Google (and historically speaking several other large private organization) is when the government steps in and stops them.

    Governments often handle corruption poorly but I can rattle of many cases where governments managed to reduce corruption on their own (ie without requiring a revolution). In many cases the source of that corruption was large private organizations.




  • If they actually did this correctly, it would be great. Whether or not it’s possible, or even desirable to eliminate all hate speech, it should be possible to minimize the harms.

    When somebody mutters some hateful comment to themselves, do we care? Not really. We care that the hateful comment gets repeated and amplified. We care that someone might take that hateful comment. We care that someone might take harmful actions based on the comment.

    If those algorithms successfully let these comments die in ignominy they’ve done their job. My fear is that they won’t really do this though. Instead they’ll mostly hide these comments and “accidentally” let them slip out when the company thinks they need an “engagement” boost.


  • Covering the second half:

    I hadn’t heard of Elsagate and had to look it up. How does AI factor into that? As near as I can tell Elsagate started with some random guy making disturbing videos and mislabeling them as child-friendly.

    I’m a good bit older than you so my nostalgia doesn’t take me lead me to any of the title you mentioned. For the most part it’s stories that aren’t covered by anyone’s IP. My childhood had a lot of folk tales recited from memory. Those stories were fairly common but there would be regional variation and most tellers would put their own twist on the stories (for example, when my Aunt told the story of the Seven Kids she would do a particular squeaky voice when she got to the part where the wolf swallows the chalk (in her version it was always chalk). That’s actually quite close to how LLMs work. She heard various versions of that story throughout her life, then she repeats it with some other bits that she incorporated from the rest of her life. I do the same thing when I retell the story to my children. It’s basically the same story my Aunt told but I translate it into English and add some modern slang.

    What would stop an AI from writing Scar into the Lion King? If you told an LLM to, “Write Hamlet but have all the royal family be Lions,” it’s likely you’d get some evil lion version of Claudius.

    There were a lot of homosexual coded villains in older media. There were also a lot of films where all the black people were bad guys, all the Asian people were goofy servants and all the women were housewives or prizes. The general consensus today is that those choices were horribly discriminatory. If AI manages to avoid that sort of behavior it would be a good thing.

    The flip side is also that artists can just as easily slip hateful material into otherwise reasonable art. Human history is full of unethical choices. Even if the AI itself doesn’t have ethics the people using it can be held to the same ethical standards as the users of any other tool or medium.