Open tcarmelveilleux opened 2 years ago
I think this is the right thing to do. I know the spec previously had prose describing that resumption information should be deleted if resumption fails (but it appears to not state this anymore). In any case, success or failure should lead to the deletion of the resumption state.
Found during internal security review at Google.
What security problem would we hope to fix by making this change? In all cases, after session establishment, we are trusting peers not to continue using a session if their credentials have changed. This is true during ongoing session communication and is equally true for session resumption. Disallowing multiple session resumption contexts for a given peer does not remove the need to trust the peer because this does not address the case where a peer holds a single resumption context and continues using this after its credentials have changed.
Edit: discussed offline. I misunderstood what was being reported here, which is not really a security problem. The sdk's resumption storage implementation just has a bug. This API implies and requires that the resumption storage delegate only store a single entry per peer, but the implementation stores more than one:
virtual CHIP_ERROR FindByScopedNodeId(const ScopedNodeId & node, ResumptionIdStorage & resumptionId,
Crypto::P256ECDHDerivedSecret & sharedSecret, CATValues & peerCATs) = 0;
However, as an amendment to the proposed solution, should we just make this happen automatically in the save() method? Surely it should not be permissible for save() to break the underlying storage. So it seems that save() must either check first for existing entries and error out for this case, or accept a duplicate and silently discard existing entries.
I'm not quite understand the role of ResumptionId
. If there can by only one resumption entry per node, why not use a stable tuple of <CompressedFabricId, NodeId>
as the ResumptionId
?
In this way, we don't need to store the relationship between the ResumptionId
and the ScopedNodeId
. I don't think it will expose any security flaws, since both CompressedFabricId
and NodeId
are public, and the shared secret sill kept secret, and correct key is needed to generate correct initiatorResumeMIC
, which can be verified by the peer.
@tcarmelveilleux @turon @bzbarsky-apple @msandstedt
That sounds like a question for @balducci-apple
Without the ResumptionID
we don't know the NodeId
of the peer that is sending us a Sigma1 with resumption
.
It doesn't mention what to do with the previous resumption state, but a careful consideration of the side-effects implies that since the resumption ID set by the responder and processed by the initiator gets updated in "session context" and that a given "session" has to be resumed, the previous resumption state, if any, for the given peer, should be deleted, prior to saving the new resumption state. Otherwise, it is possible that a session be multiply resumed, or that resumption contexts that are now logically expired, be reused.
The current implementation of DefaultSessionResumptionStorage never removes a prior entry for a given peer, except when fabric removal is called.
Proposed Solution Either:
Delete(ScopedNodeId) should always be called before Save(...) Save(...) should automatically called Delete (and ignore "failure/missing") for the ScopedNodeId of the peer
Furthermore, Delete(ScopedNodeId) must support trying to delete any possible duplicate within the index, if any exists (i.e. make sure there remains no reume contexts). This makes sense since there should only need to be one node resumption state between peers.
One solution to achieve the proactive cleaning for individual implementations is to override the default session resumption storage implementation. But if we would like this in the default implementation itself, code needs to be added or changed.
Problem
Found during internal security review at Google.
In either of these cases, a resumption has successfully occurred:
The spec says:
It doesn't mention what to do with the previous resumption state, but a careful consideration of the side-effects implies that since the resumption ID set by the responder and processed by the initiator gets updated in "session context" and that a given "session" has to be resumed, the previous resumption state, if any, for the given peer, should be deleted, prior to saving the new resumption state. Otherwise, it is possible that a session be multiply resumed, or that resumption contexts that are now logically expired, be reused.
The current implementation of
DefaultSessionResumptionStorage
never removes a prior entry for a given peer, except when fabric removal is called.Proposed Solution
Either:
Delete(ScopedNodeId)
should always be called beforeSave(...)
Save(...)
should automatically calledDelete
(and ignore "failure/missing") for the ScopedNodeId of the peerFurthermore,
Delete(ScopedNodeId)
must support trying to delete any possible duplicate within the index, if any exists (i.e. make sure there remains no reume contexts). This makes sense since there should only need to be one node resumption state between peers.