• 1 Post
  • 856 Comments
Joined 3 years ago
cake
Cake day: June 18th, 2023

help-circle







  • A good forum design will only get you so far, the rest is up to the moderators. If you let bad actors in, it doesn’t matter how you designed your forum, they will poison the well and drive other people out.

    Yes, well, the problem with hexbear was that it started with bad actors. As they made their true colors apparent to the lemmy community at large, they were increasingly defederated.

    The best communities I’ve been in are in independent old-style forums. One of them is Tildes. Most of these don’t feature downvotes (or upvotes for that matter) and are honestly the better places to have discussions IMO.

    Oh yes, my past experience is in old web forums as well. Those communities were more isolated though, they essentially existed inside their own bubbles. Unregistered users could read them, but not participate in any functional way, and typically the people that found them were looking for a community like that on purpose.

    I’ve also experienced such communities becoming toxic due to the actions of individual moderators or admins, post voting not required.

    Ultimately I think I agree this far - if you’re going to disable voting you should do all of it. Removing only the downvoting is the YouTube path, the authoritarian path, the toxic positivity path.


  • NaibofTabr@infosec.pubtoMemes@sopuli.xyzYou earned some more dislikes
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    1
    ·
    14 days ago

    I wouldn’t imply that the unhealthyness of hexbear is due to the disabled downvoting though, and I’m sure you aren’t either, but just to be clear.

    No, I’m implying the reverse: disabling downvoting is a symptom of the unhealthy mentality of the people running that server. Disabling downvoting appeals to authoritarians - that is, the type of people who are interested in silencing dissent (QED).



  • Enforced toxic positivity does not produce better conversations or better communities. It basically just turns a discussion forum into Disneyland, where everyone is happy all the time, because there’s no other option. It’s the kind of yes-man thinking you get in corporate meetings that produce really bad ideas because “don’t be negative! there are no bad ideas here!”






  • My guess is mechanical stress during the print.

    Think of your print object as a lever. The attachment to the print bed is the fulcrum. The taller the object gets, the longer the lever arm and the more potential for movement, especially while the plastic is still warm and soft.

    On the other end of the lever is the nozzle spitting out melted plastic. The melted plastic is sticky (PETG in particular is kind of like chewing gum at print temperature). As the nozzle moves across the printed surface, the sticky plastic pulls on the previous layer, exerting a lateral force (you can watch this happen during the print, it’s most obvious with tall thin parts). If there isn’t enough contact area between the topmost layer and the one below it (which in your case it appears those parts of the hexagons have very little contact with the layer below) then the top layer can be ripped off.

    Basically the individual limbs of the hexagons are too thin, and the angles are too steep. As the print gets taller the whole thing will flex more, making failures more likely near the top.




  • AI coding tools can do common, simple functions reasonably well, because there are lots of examples of those to steal from real programmers on the Internet. There is a large corpus of data to train with.

    AI coding tools can’t do sophisticated, specific-case solutions very well, because there aren’t many examples of those for any given use case to steal from real programmers on the Internet. There is a small corpus of data to train with.

    AI coding tools can’t solve new problems at all, because there are no examples of those to steal from real programmers on the Internet. There is no corpus of data to train with.

    AI coding tools have already ingested all of the code available on the Internet to train with. There is no more new data to feed in. AI coding tools will not get substantially better than they are now. All of the theft that could be committed has been committed, which is why the AI development companies are attempting to feed generated training material into their models. Every review of this shows that it makes the output from generative models worse rather than better.

    Programming is not about writing code. That is what a manager thinks.
    Programming is about solving problems. Generative AI doesn’t think, so it cannot solve problems. All it can do is regurgitate material that it has previously ingested which is hopefully close-ish to the problem you’re trying to solve at the moment - material which was written by a real thinking human that solved that problem (or a similar one) at some point in the past.

    If you patronize a generative AI system like Claude Code, you are paying into, participating in, and complicit in, the largest example of labor theft in history.