• 0 Posts
  • 134 Comments
Joined 11 months ago
cake
Cake day: July 28th, 2023

help-circle
  • This happens to me constantly. Just the other day I asked some friends for something and then they sent the literal exact opposite of that thing. Pretend I asked for blue with red stripes they gave me green with yellow polka dots. And it wasn’t just one person it was three separate people who all decided that made sense for some reason.

    I was extremely specific too, even more than usual because I know people constantly misinterpret me. I made extra sure to not use any language with vague meanings and it still happened anyway. It’s like we live in alternate realities where words have completely different meanings.

    It makes me not want to talk to people at all.



  • Sorry, I misinterpreted what you meant. You said “any AI models” so I thought you were talking about the model itself should somehow know where the data came from. Obviously the companies training the models can catalog their data sources.

    But besides that, if you work on AI you should know better than anyone that removing training data is counter to the goal of fixing overfitting. You need more data to make the model more generalized. All you’d be doing is making it more likely to reproduce existing material because it has less to work off of. That’s worse for everyone.


  • What you’re asking for is literally impossible.

    A neural network is basically nothing more than a set of weights. If one word makes a weight go up by 0.0001 and then another word makes it go down by 0.0001, and you do that billions of times for billions of weights, how do you determine what in the data created those weights? Every single thing that’s in the training data had some kind of effect on everything else.

    It’s like combining billions of buckets of water together in a pool and then taking out 1 cup from that and trying to figure out which buckets contributed to that cup. It doesn’t make any sense.



  • ayaya@lemdro.idtoLinux@lemmy.mlFlathub has passed 2 billion downloads
    link
    fedilink
    English
    arrow-up
    35
    arrow-down
    4
    ·
    edit-2
    8 days ago

    For me on Arch, Flatpaks are kinda useless. I can maybe see the appeal for other distros but Arch already has up-to-date versions of everything and anything that’s missing from the main repos is in the AUR.

    I also don’t like how it’s a separate package manager, they take up more space, and to run things from the CLI it’s flatpak run com.website.Something instead of just something. It’s super cumbersome compared to using normal packages.


  • ayaya@lemdro.idtolinuxmemes@lemmy.worldlow effort maymay
    link
    fedilink
    English
    arrow-up
    13
    ·
    9 days ago

    Same here. Switched to Arch in 2015 so I am also coming up on the 9 year mark. I have had very few issues, and the ones I have had were usually my fault for doing something stupid. I used Windows, OS X, and Ubuntu previously and compared to those Arch is a dream. Hence why I’ve stuck with it for so long now.





  • I pretty much never reboot the Pi. It currently has over 18 months of uptime on it. My NAS on the other hand I probably restart for one reason or another maybe once every 6 months. So yeah I’d say I reboot it minimum 3x more often.

    Plus a reboot takes much longer on my NAS than on the Pi. The server board is slow to start, the SAS cards are slow to start, and unRAID is slow to start. Then I need to manually enter the password for disk encryption. Then wait for the array to start up. Then wait a bit more for the docker containers to start. Add all of that up and even the absolute fastest reboot is like 10 minutes while the Pi probably takes 30 seconds.

    And what if I want to swap hard drives? Now it’s down for an hour. I guess I could wait until 3am to do all my upgrades so everyone is asleep, but I’d rather not. I suppose if it were just for myself it would matter a lot less. But again, it’s only $15 to not have to think about it at all.