• 21 Posts
  • 160 Comments
Joined 1 year ago
cake
Cake day: June 19th, 2023

help-circle
  • Thanks for sharing this. I wasn’t familiar with this channel, not I’m liking it.

    I just read that this guy was part of Nebula and was forced out. It’s remarkable that he’s forced out for speaking openly and defending his beliefs when Isaac Arthur is tolerated despite having much more onerous politics but having them in secret. Smh.


  • Andy@slrpnk.netOPtoMemes@lemmy.mlMastermind
    link
    fedilink
    arrow-up
    1
    ·
    edit-2
    3 days ago

    Amen. It drives me fuckin’ nuts anytime – in business as well as in sci-fi and general discussion – when people envision a society made perfect because it’s run by a genius computer.

    For pretty much every challenge society faces, the major obstacle is not that we’re unsure what to do or lack the intelligence to solve. We already have all the solutions, it’s just that our decision making systems are completely disinterested in employing any of the solutions that we already have.

    It’s like, if you could get everyone to agree to listen to a computer, why not just skip the computer and get everyone to agree to listen to a combination of popular will and expert advice? Popular will and expert advice are like the supercomputer that runs society that we already have.




  • This article doesn’t really answer most of my questions.

    What subjects does the AI cover? Do they do all their learning independently? Does AI compose the entire lesson plan? What is the software platform? Who developed it? Is this just an LLM or is there more to it? How are students assessed? How long has the school been around, and what is their reputation? What is the fundamental goal of their approach?

    Overall, this sounds quite dumb. Just incredibly and transparently stupid. Like, if they insisted that all learning would be done on the blockchain. I’m very open minded, but I don’t understand what the student’s experience will be. Maybe they’ll learn in the same way one could learn by browsing Wikipedia for 7 hours a day. But will they enjoy it? Will it help them find career fulfillment, or build confidence or learn social skills? It just sounds so much like that Willie Wonka experience scam but applied to an expensive private school instead of a pop-up attraction.


  • I feel like this is a pretty crass joke to make.

    A good friend of mine found a body a few months ago. It’s a pretty shitty experience. And it’s actually a lot like what OP describes. A sense of foreboding and suspicion combined with a conviction that these thoughts are foolish. And an uncertainty whether to check or to alert someone or to just try to forget it.

    Op, I’d report it and ask them to please follow up with you and let you know. It’s probably nothing, and you’ll feel better once you know it was nothing, and that you did the responsible thing in having it dealt with.







  • I will also add that I think in the long run, as we try to figure out how to differentiate between humans and machines, the only real reliably solution I see is to focus on elevating the individual. Having people with long histories validate their reality by living and documenting it.

    I don’t upvote something that I’d be ashamed for someone to see I upvote. I might make an exception for pornographic content, but even with that, if it’s pseudononymous in that it’s not attached to my personal public life, I don’t mind if someone can trace through and see what a specific account I use for those purposes has liked and disliked.


  • I don’t think it’s secret. A lot of OpenAI’s business strategy is to warn of the danger of their own project as a means of hyping it.

    OpenAI, despite having produced a pretty novel product, doesn’t really have a sound business model. LLMs are actually expensive to run. The energy and processing is not cheap, and it’s really not clear that they produce something of value. It’s a cool party trick, but a lot of the use cases just aren’t cost effective at this point. That makes their innovation hard to commercialize. So OpenAI promotes itself like online clickbait games.

    You know the ones that are like, ‘WARNING: This game is so sexy it is ADDICTIVE! Do NOT play our game if you don’t want to CUM TOO HARD!’

    That’s OpenAI’s marketing strategy.








  • Haha a bike.

    I hold out hope, actually, that as the right-to-repair movement continues to grow, eventually repairability and control will become more common consumer interests, in the same way that vehicle safety wasn’t something people thought about when buying a car before the 70s, and now it’s one of the main influences when buying a car.

    Once people start caring – and again, I believe this is the direction we’re heading – it will become something manufacturers have to design for.


  • This is modestly interesting. My brother worked here before they had layoffs about two years ago, and had a generally favorable opinion of the company and leadership.

    Fundamentally, while I think RJ seems like a sound businessman and technologist, and I like the company’s taste a bit, I will never be able to reconcile his views with mine. He very openly views cars as computers and software and services that happen to move you around, and I would like it to be a machine over which I have as minimal a relationship as possible with the manufacturer after I acquire the product.

    Still, I wish them luck.


  • I agree with you… but I don’t disagree with @john_mcmurray.

    The fact that Biden actually participates in genocide at all is just ghastly. I honestly struggle to find the words for it. And once he crossed that bridge, I will never ever try to argue with someone who assigns him equal moral equivalence with Trump.

    It’s like debating whether someone who killed one of your children is better than someone who killed two. Mathematically? I think so. Conceptually? Those are who child murderers. The difference is negligible.

    There are still Palestinians alive right now, though, so we can’t give up, no matter how heartbroken.