> > Overall, it seems like having clients connect *directly* to the > > snapshot volumes once they've been started might have avoided some > > complexity or problems. Was this considered? > > Can you explain this in more detail? Are you saying that the virtual > namespace overlay used by the current design can be reused along with > returning extra info to clients or is this a new approach where you > make the clients much more intelligent than they are in the current > approach? Basically the clients would have the same intelligence that now resides in snapview-server. Instead of spinning up a new protocol/client to talk to a new snapview-server, they'd send a single RPC to start the snapshot brick daemons, then connect to those itself. Of course, this exacerbates the problem with dynamically changing translator graphs on the client side, because now they dynamically added parts will be whole trees (corresponding to whole volfiles) instead of single protocol/client translators. Long term, I think we should consider *not* handling these overlays as modifications to the main translator graph, but instead allowing multiple translator graphs to be active in the glusterfs process concurrently. For example, this greatly simplifies the question of how to deal with a graph change after we've added several overlays. * "Splice" method: graph comparisons must be enhanced to ignore the overlays, overlays must be re-added after the graph switch takes place, etc. * "Multiple graph" method: just change the main graph (the one that's rooted at mount/fuse) and leave the others alone. Stray thought: does any of this break when we're in an NFS or Samba daemon instead of a native-mount glusterfs daemon? > > If we're using either on-disk or on-network encryption, client keys > > and certificates must remain on the clients. They must not be on > > servers. If the volumes are being proxied through snapview-server, > > it needs those credentials, but letting it have them defeats both > > security mechanisms. > > > > Also, do we need to handle the case where the credentials have > > changed since the snapshot was taken? This is probably a more > > general problem with snapshots themselves, but still needs to be > > considered. > > Agreed. Very nice point you brought up. We will need to think a bit > more on this Jeff. This is what reviews are for. ;) Another thought: are there any interesting security implications because USS allows one user to expose *other users'* previous versions through the automatically mounted snapshot? _______________________________________________ Gluster-devel mailing list Gluster-devel@xxxxxxxxxxx http://supercolony.gluster.org/mailman/listinfo/gluster-devel