![](/static/253f0d9b/assets/icons/icon-96x96.png)
![](https://lemmy.world/pictrs/image/8f2046ae-5d2e-495f-b467-f7b14ccb4152.png)
I’d rather have “star” or “approval” voting, but I’ll take ranked choice voting over what we’ve currently got.
I’d rather have “star” or “approval” voting, but I’ll take ranked choice voting over what we’ve currently got.
I think the consequences will be different for every case and based on what was done.
This ruling is a major step forward so that cases can actually be brought against corrupt politicians who are abusing their power in this way.
In this case, Sylvia was falsely accused, arrested, and then the charges were dropped a day later.
That’s different when compared to Trump where he went through the court system and was found guilty on all counts.
Link to the court’s opinion, if anyone is interested in reading through it:
https://www.supremecourt.gov/opinions/23pdf/22-1025_1a72.pdf
So, roughly a radius of ~415 Miles (667km).
Fixed it!
Thanks! It’s a good read and I like the idea of a private cloud compute (PCC) system, but that doesn’t mention anywhere that ChatGPT will be running in that PCC system (if you were trying to imply that).
And while OpenAI could implement something similar to PCC, I haven’t seen them announce that anywhere either.
I’d say the proof is on Apple to show that it’s being done on-device or that all processing is done on iCloud servers.
You’re saying that OpenAI is just going to hand over their full ChatGPT model for Apple to set up on their own servers for free?
But from the article itself:
the partnership could burn extra money for OpenAI, because it pays Microsoft to host ChatGPT’s capabilities on its Azure cloud
I get it if they created a small version of their LLM to run locally, but I would expect Apple to pay a price even for that.
I think you may be confusing this ChatGPT integration with Apple’s own LLM that they’re working on… Again, from the linked article:
Still, Apple’s choice of ChatGPT as Apple’s first external AI integration has led to widespread misunderstanding, especially since Apple buried the lede about its own in-house LLM technology that powers its new “Apple Intelligence” platform.
As long as they bring some whiskey and rye
What? No. I would rather use my own local LLM where the data never leaves my device. And if I had to submit anything to ChatGPT I would want it anonymized as much as possible.
Is Apple doing the right thing? Hard to say, any answer here will just be an opinion. There are pros and cons to this decision and that’s up to the end user to decide if the benefits of using ChatGPT are worth the cost of their data. I can see some useful use cases for this tech, and I don’t blame Apple for wanting to strike while the iron is hot.
There’s not much you can really do to strip out identifying data from prompts/requests made to ChatGPT. Any anonymization of that part of the data is on OpenAI to handle.
Apple can obfuscate which user is asking for what as well as specific location data, but if I’m using the LLM and I tell it to write up a report while including my full name in my prompt/request… that’s all going directly into OpenAIs servers and logs which they can eventually use to help refine/retrain their model at some point.
I’m sure you understand this, but anonymized data doesn’t mean it can’t be deanonymized. Given the right kind of data, or enough context they can figure out who you are fairly quickly.
Ex: You could “Anonymize” gps traces, but it would still show the house you live at and where you work unless you strip out a lot of the info.
http://androidpolice.com/strava-heatmaps-location-identity-doxxing-problem/
Now with LLMs, sure, you could “anonymize” which user said or asked for what… but if something identifying is sent in the request itself, it won’t be hard to deanonymize that data.
Thanks, I wandered in from “All”. Content looks fun in here so I’ll subscribe and stick around for a while.
I still need to watch the latest Netflix adaptation, of the show, it just hasn’t been a high priority for me yet.
I feel out of the loop on this one.
Did Netflix have a publicity fail recently where they argued about how posting some flyers in a common area for employees was enough to satisfy some law?
(All I get from searching “Netflix” and “Blind” are news results for “Love is Blind”).
Well, it won’t be easy for Florida to go hang with Hawaii and California.
*swell
They’re gonna have to build a wall to keep the water out!
I thought I read something earlier today that mentioned that the current deal is that the hostages be returned and then Israel will pull out it’s troops.
Hamas wants to alter it so that they release a few hostages, then Israel pulls out, and then they release the rest of the hostages.
It sounds like Israel had already agreed to that deal so isn’t this in Hamas’ ball park to accept or not?
Then again it sounds like every time the deal is altered, the other side wants to male new changes,
As Blinken mentioned in a different article:
“At some point in a negotiation, and this has gone back and forth for a long time, you get to a point where if one side continues to change its demands, including making demands and insisting on changes for things that it already accepted, you have to question whether they’re proceeding in good faith or not.” https://apnews.com/article/israel-palestinians-hamas-war-ceasefire-negotiations-ab6925549d8f523a6e5c61e88e7eec8a
Has anyone actually published the entire deal (at any stage) in its entirety so that everyone can see what is being debated?
Corporations are big into lobbying. “Studies” like this help them to convince lawmakers to make decisions that benefit them.
In this case, they might not be able to easily lower minimum wage, but they can say that it’s been a burden and try to get a break in other ways.
Edit in response to the edited comment above:
Poe’s Law, should have included a “/s” at least.
https://en.m.wikipedia.org/wiki/Poe’s_law
AI can be very useful, the problem here is humans trusting it to be accurate all the time.
In this case it should be used to narrow down results, but even then the police need to do their job. They need to do an actual investigation to gather evidence that they have the right person before even attempting to make am arrest.
Even removing AI from the picture entirely doesn’t solve this problem. Just look at wrongful arrests that have been made simply because a wanted criminal has the exact same name as someone else.