In this case, we've successfully pushed back with the granting agency (US NIH,
generally, for us) that it's just not feasible to guarantee that the data
are truly gone on a production parallel filesystem. The data are encrypted
at rest (including offsite backups), which has been sufficient for our
purposes. We'll then just use something like GNU shred(1) to do a
best-effort secure delete.

In addition to RAID, other confounding factors to be aware of are snapshots
and cached data.

On Wed, Sep 29, 2021 at 10:52:33AM -0400, Paul Edmon via Beowulf wrote:
> I guess the question is for a parallel filesystem how do you make sure you
> have 0'd out the file with out borking the whole filesystem since you are
> spread over a RAID set and could be spread over multiple hosts.
> 
> -Paul Edmon-
> 
> On 9/29/2021 10:32 AM, Scott Atchley wrote:
> > For our users that have sensitive data, we keep it encrypted at rest and
> > in movement.
> > 
> > For HDD-based systems, you can perform a secure erase per NIST
> > standards. For SSD-based systems, the extra writes from the secure erase
> > will contribute to the wear on the drives and possibly their eventually
> > wearing out. Most SSDs provide an option to mark blocks as zero without
> > having to write the zeroes. I do not think that it is exposed up to the
> > PFS layer (Lustre, GPFS, Ceph, NFS) and is only available at the ext4 or
> > XFS layer.
> > 
> > On Wed, Sep 29, 2021 at 10:15 AM Paul Edmon <ped...@cfa.harvard.edu
> > <mailto:ped...@cfa.harvard.edu>> wrote:
> > 
> >     The former.  We are curious how to selectively delete data from a
> >     parallel filesystem.  For example we commonly use Lustre, ceph,
> >     and Isilon in our environment.  That said if other types allow for
> >     easier destruction of selective data we would be interested in
> >     hearing about it.
> > 
> >     -Paul Edmon-
> > 
> >     On 9/29/2021 10:06 AM, Scott Atchley wrote:
> > >     Are you asking about selectively deleting data from a parallel
> > >     file system (PFS) or destroying drives after removal from the
> > >     system either due to failure or system decommissioning?
> > > 
> > >     For the latter, DOE does not allow us to send any non-volatile
> > >     media offsite once it has had user data on it. When we are done
> > >     with drives, we have a very big shredder.
> > > 
> > >     On Wed, Sep 29, 2021 at 9:59 AM Paul Edmon via Beowulf
> > >     <beowulf@beowulf.org <mailto:beowulf@beowulf.org>> wrote:
> > > 
> > >         Occassionally we get DUA (Data Use Agreement) requests for
> > >         sensitive
> > >         data that require data destruction (e.g. NIST 800-88). We've
> > >         been
> > >         struggling with how to handle this in an era of distributed
> > >         filesystems
> > >         and disks.  We were curious how other people handle requests
> > >         like this?
> > >         What types of filesystems to people generally use for this
> > >         and how do
> > >         people ensure destruction?  Do these types of DUA's preclude
> > >         certain
> > >         storage technologies from consideration or are there creative
> > >         ways to
> > >         comply using more common scalable filesystems?
> > > 
> > >         Thanks in advance for the info.
> > > 
> > >         -Paul Edmon-
> > > 
> > >         _______________________________________________
> > >         Beowulf mailing list, Beowulf@beowulf.org
> > >         <mailto:Beowulf@beowulf.org> sponsored by Penguin Computing
> > >         To change your subscription (digest mode or unsubscribe)
> > >         visit https://beowulf.org/cgi-bin/mailman/listinfo/beowulf
> > >         <https://beowulf.org/cgi-bin/mailman/listinfo/beowulf>
> > > 

> _______________________________________________
> Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
> To change your subscription (digest mode or unsubscribe) visit 
> https://beowulf.org/cgi-bin/mailman/listinfo/beowulf


-- 
Skylar
_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit 
https://beowulf.org/cgi-bin/mailman/listinfo/beowulf

Reply via email to