• 0 Posts
  • 26 Comments
Joined 1 year ago
cake
Cake day: June 5th, 2023

help-circle
  • Just like the shopping cart theory itself, this is mostly just a thought experiment at this point in time.

    The point of a protestation is to make it hard for others to ignore, and make it clear what the end condition is. I don’t plan on just starting to do this as an individual because it would have no impact; I still make sure my own carts get returned personally.

    The point stands that our goodwill is frequently exploited for profit, often under the pretense that it’s just basic human decency.


  • I gotta be honest, my bf or I still make sure the cart goes back every time we shop, but I increasingly question whether I should bother. These grocery stores keep raising their prices well above inflation so they can pocket the rest and brag to shareholders about it, at the cost of people who actually shop there.

    It’s tempting to say that if they’re going to play that game, they get no courtesy from me as a customer and can hire more cart collectors. It’s miniscule on an individual level, but it is unpaid labor.

    There’s the argument that unreturned carts mostly inconvenience other customers, but honestly if the store is exploiting both customers’ goodwill and wallets, I think it’s fine to make the experience at that store just that little bit worse; maybe that last little push will encourage people to shop elsewhere (where it’s an option of course, i.e. not a small town).

    I don’t feel this urge at stores like H Mart even though they have so many fewer return stalls and it’s often a longer walk to do so.

    I guess this is kind of an antithesis to Shopping Cart Theory I’ve been developing in my head over the past little while. It’s conditional on the store itself being overtly greedy, but I think there might be something to it.



  • I’d also consider myself pretty tech-savvy, but that came from plenty of mistakes growing up including putting malware on the family computer at least twice (mostly ads for these “Pokemon MMOs” back in the mid aughts that were too enticing for my kid brain to refuse 😅).

    It’s very easy for me to forget how much of an outlier my tech experience is among most folks around my age. I had an acquaintance in the first year of college I helped by giving essay advice, and was very surprised to see that the only thing they really knew how to do was basic use of apps on their iPhone. They got a laptop for school, but no computer experience, no keyboard typing experience, and even just the iPhone Settings app was a scary place to be avoided for the most part. To this person, Microsoft Word was a new thing they had to learn on top of everything else. In college. It was also in the South so I don’t know if I should be that surprised unfortunately.

    Regardless, it was pretty wild to me, but a very real reminder that not everyone has access to the same resources education, and/or experience to draw on.






  • >dinner is done
    >announce dinner
    >everyone shows up 10 minutes later

    Some of the food is cold because I didn’t cover it, but why should I? It would have been fine if they came to the table when I announced dinner.

    >next day
    >dinner is almost done
    >remembering yesterday, I decide to announce dinner 10 minutes early so that 10 minutes later is “on time”
    >everyone arrives to the table immediately, remembering that it was cold yesterday not wanting that to happen again
    >mfw




  • The device wouldn’t necessarily have to be constantly streaming the audio to a central server. If it’s capable of hearing wake up words like “Ok Google” it’s capable of listening for other phrases and having onboard processing to relay back the results much more compressed. Whether or not this is common practice is another matter, and yes the algorithms are scary good even without eavesdropping.





  • That’s fair. I think fundamentally a false positive/negative isn’t that much different. Pretty much all tests—especially those dealing with real world conditions—are heuristic, as are all LLMs by necessity of the design. Hallucination is a pretty specific term given to AI as an attempt to assign agency to a system that doesn’t actually have any (by implying it’s crazy and making stuff up instead of a black box with deterministic inputs and outputs spitting out something factually wrong but with a similar format to what is trained on). I feel like the nature of any tool where “you can’t trust this to be entirely accurate” should have an umbrella term that encompasses both types of providing inaccurate info under certain conditions.

    I suppose the difference is that AI is a lot more likely to randomly go off, whereas a blood test is likelier to provide repeated false positives for the same person with their unique biology? There’s also the fact that most medical tests represent a true/false dichotomy or lookup table, whereas an LLM is given the entire bounds of language.

    Would an AI clustering algorithm (say, K-means for instance) giving an inaccurate diagnosis be a false positive/negative or a hallucination? These models can be programmed on a sliding scale and I feel like there’s definitely an area where the line could get pretty blurry.



  • I mean, AI is used in fraud detection pretty often; when it hits a false positive (which happens frequently on a population-level basis), is that not a hallucination of some sort? Obviously LLMs can go off the rails much further because it’s readable text, but any machine learning model will occasionally spit out really bad guesses almost any person could have done better with. (To be fair, humans are highly capable of really bad guesses too).