• sus
    link
    fedilink
    arrow-up
    44
    arrow-down
    1
    ·
    edit-2
    il y a 3 jours

    their given reasons are “to keep backups” and “academic and clinical research with de-identified datasets”

    they seem to actually do a fairly good job with anonymizing the research datasets, unlike most “anonymized research data”, though for the raw data stored on their servers, they do not seem to use encryption properly and their security model is “the cloud hoster wouldn’t spy on the data right?” (hint: their data is stored on american servers, so the american authorities can just subpoena Amazon Web Services directly, bypassing all their “privacy guarantees”. (the replacement for the EU-US Privacy Shield seems to be on very uncertain legal grounds, and that was before the election))

    • ballmerpeaking
      link
      fedilink
      arrow-up
      6
      ·
      edit-2
      il y a 2 jours

      De-identified data is an oxymoron. Basically any dataset that’s in any way interesting is identifiable.

      • sus
        link
        fedilink
        arrow-up
        4
        ·
        edit-2
        il y a 2 jours

        no it’s not. If you reduce the information in the datapoints until none of them are unique, then it is very obviously impossible to uniquely identify someone from them. And when you have millions of users the data can definitely still be kept interesting

        (though there’s pretty big pitfalls here, as their report seems to leave open the possibility of not doing it correctly)