Closed Bug 1294095 Opened 8 years ago Closed 8 years ago

Crash in mozilla::DataChannelConnection::SctpDtlsOutput

Categories

(Core :: WebRTC: Networking, defect, P1)

49 Branch
x86
Windows
defect

Tracking

()

VERIFIED FIXED
mozilla51
Tracking Status
firefox48 --- wontfix
firefox49 + verified
firefox-esr45 49+ verified
firefox50 + verified
firefox51 + verified

People

(Reporter: philipp, Assigned: jesup)

References

Details

(Keywords: crash, regression, sec-critical, Whiteboard: [adv-main49+][adv-esr45.4+])

Crash Data

Attachments

(2 files, 1 obsolete file)

This bug was filed from the Socorro interface and is 
report bp-c237a768-7c3a-4c28-b18c-ad34f2160810.
=============================================================
Crashing Thread (54)
Frame 	Module 	Signature 	Source
0 	xul.dll 	mozilla::DataChannelConnection::SctpDtlsOutput(void*, void*, unsigned int, unsigned char, unsigned char) 	netwerk/sctp/datachannel/DataChannel.cpp:691
1 	xul.dll 	sctp_lowlevel_chunk_output 	netwerk/sctp/src/netinet/sctp_output.c:5010
2 	mozglue.dll 	je_malloc 	memory/mozjemalloc/jemalloc.c:6184
3 	xul.dll 	sctp_abort_an_association 	netwerk/sctp/src/netinet/sctputil.c:4212
4 	xul.dll 	sctp_threshold_management 	netwerk/sctp/src/netinet/sctp_timer.c:167
5 	xul.dll 	sctp_t3rxt_timer 	netwerk/sctp/src/netinet/sctp_timer.c:832
6 	xul.dll 	sctp_timeout_handler 	netwerk/sctp/src/netinet/sctputil.c:1706
7 	xul.dll 	sctp_handle_tick 	netwerk/sctp/src/netinet/sctp_callout.c:155
8 	xul.dll 	user_sctp_timer_iterate 	netwerk/sctp/src/netinet/sctp_callout.c:194
9 	kernel32.dll 	BaseThreadInitThunk 	
10 	ntdll.dll 	__RtlUserThreadStart 	
11 	ntdll.dll 	_RtlUserThreadStart

crashes with this signature are regressing in number since the 49 nightly cycle & the signature is currently making up 0.3% of browser crashes in 49.0b1.
e5e5 signature

I wonder if this is another instance of the sctp timer iterator issue?  Byron?

I don't see how peer could point to freed memory, or peer->mSTS, while the sctp stack is still open.  In DataChannelConnection::Destory() we close() the sctp socket on STS thread (mSTS), and the runnable that does that has a RefPtr to the DataChannelConnection, so we can't free the DataChannelConnection until we've finished the close().  This implies that usrsctp_close() could leave a dangling timer reference about to fire...
Assignee: nobody → rjesup
Group: media-core-security
Rank: 5
Flags: needinfo?(tuexen)
Flags: needinfo?(docfaraday)
Keywords: sec-critical
Priority: -- → P1
I don't really see how this pointer can go wrong... Are you also tearing down the SCTP stack at the same time?
Flags: needinfo?(tuexen)
So we're not calling sctp_finish until the browser is shutdown, so the timer thread will continue to operate.  However, as per above, we are certain that until usrsctp_close() returns that the pointer will be valid.  I would hope we would not get an callback to send a packet (per the stack) after close() returns...
(In reply to Michael Tüxen from comment #2)
> I don't really see how this pointer can go wrong... Are you also tearing
> down the SCTP stack at the same time?

The stack is not being torn down, but we are calling |usrsctp_deregister_address| and |usrsctp_close|. Should that be enough?
Flags: needinfo?(docfaraday)
I think so.
(In reply to Byron Campen [:bwc] from comment #4)
> (In reply to Michael Tüxen from comment #2)
> > I don't really see how this pointer can go wrong... Are you also tearing
> > down the SCTP stack at the same time?
> 
> The stack is not being torn down, but we are calling
> |usrsctp_deregister_address| and |usrsctp_close|. Should that be enough?

Note: the order of usrsctp_deregister_address and usrsctp_close() isn't fixed here (and not on the same thread).  Could that impact it?  We could always call deregister_address before usrsctp_close though.  Calling it afterwards would be possible; we'd have to pass a refptr to the connection along with the DestroyOnSTS(), and move the deregister into DestroyOnSTS (along with the usrsctp_close()es).
When looking at the traces at https://crash-stats.mozilla.com/signature/?product=Firefox&version=49.0b&signature=mozilla%3A%3ADataChannelConnection%3A%3ASctpDtlsOutput&_columns=date&_columns=product&_columns=version&_columns=build_id&_columns=platform&_columns=reason&_columns=address&_sort=-date&page=1
it is
* always Windows as a platform
* always when a timer (either HEARTBEAT or data retransmission timer) runs off
* always when this results in aborting the association due to excessive retransmissions.
So how do you think usrsctp_close() is in the game here?
A question: When do you call usrsctp_deregister_address()? When the Peer connection dies? When it get closed? Do you leave some SCTP associations open?
Based on that I could do some testing with the usrsctp stack using valgrind, that should give a hint if the current code handles the situation well. This wouldn't mean that the code you are using, handles it, but possibly I can identify a bug in the current code...
So the only way the DataChannelConnection object can be freed is if we go through Destroy() first, which fires off DestroyOnSTS (which does usrsctp_close()) to the STS thread, then calls deregister_address (on the main/calling thread).  Thus due to thread ordering the DestroyOnSTS may run before or after deregister_address.  I presume the crashes are due to a race between the abort timer going off and the teardown of the peerconnection.

After Destroy() returns, the DataChannelConnection object will be released, which (if that's the last ref) will delete it.  (Various things in-flight might hold a ref to the connection, but generally it will be freed on return from Destroy()).

This pattern assumes that after deregister_address, no callbacks can occur (the DataChannelConnection ptr (raw ptr) is associated with the socket callback).  If instead we have to do usrsctp_close to stop callbacks, or both that and deregister, we could move deregister to before we fire off DestroyOnSTS(), or if deregister must come after the close, move it to DestroyOnSTS and also pass a reference to DataChannelConnection to DestroyOnSTS, so that the connection object can't be deleted until DestroyOnSTS completes.
Flags: needinfo?(tuexen)
Let me test a bit...
Flags: needinfo?(tuexen)
This signature is still on the rise for 49 beta 2 & 3, though not huge volume. Tracking.
So, looking through the stuff that |usrsctp_close| calls, I don't see an |sctp_timer_stop| for |SCTP_TIMER_TYPE_SEND|, which is the timer type that is blowing us up here. I only see an |sctp_timer_stop| for |SCTP_TIMER_TYPE_NEWCOOKIE|. Is there something else that |usrsctp_close| does that ought to be stopping (or at least neutering) this timer?
Flags: needinfo?(tuexen)
It is for example stopped when we clear the net in the sctp_free_remote_addr() macro.

My current guess is that it is fine that the timer fires. The problem is that as a result
of the firing, the upper layer gets notified and probably deregisters the pointer to the lower
layer. I need to understand the C++ stuff...
I can add code that the lower layer pointer is not provided anymore after it is deregistered,
that might solve the issue. But I would like to understand the issue first. Not sure if I
can reproduce the problem with firefox... I was busy with my dayjob earlier this week...
Flags: needinfo?(tuexen)
Proposed wrapper idea, though Michael's idea may be better
Attachment #8782576 - Flags: feedback?(docfaraday)
OK, I looked at the code. It is not so easy as I thought to add the protection. And I have no
idea why this is happening.

If you just want to avoid the problem to occur, add

	usrsctp_sysctl_set_sctp_assoc_rtx_max_default(0xffffffff);

after
https://dxr.mozilla.org/mozilla-central/source/netwerk/sctp/datachannel/DataChannel.cpp#351

This will basically turn off the detection that the peer is dead. All error logs I looked at, this was the code path the problem occurs.

This doesn't solve the real problem (which I currently haven't been to identify based on the information available) but will avoid the problem.
Flags: needinfo?(rjesup)
Comment on attachment 8782576 [details] [diff] [review]
wrap and lock DataChannelConnection for SCTP callbacks

Review of attachment 8782576 [details] [diff] [review]:
-----------------------------------------------------------------

::: netwerk/sctp/datachannel/DataChannel.h
@@ +296,5 @@
>    nsCOMPtr<nsIThread> mInternalIOThread;
> +
> +  // Temporary wrapper object to use with sctp library callbacks so we can
> +  // neuter it on Destroy().  Leaked on purpose when we delete this object.
> +  ConnectionWrapper *mTempWrapper;

Oof. Leaking a mutex on purpose? Surely there's a better fix.
Attachment #8782576 - Flags: feedback?(docfaraday) → feedback-
Hmmm.  That will turn off detection of a failed connection.  In theory ICE Consent Freshness should eventually kick in and mark it as failed.  There might even be advantages in not having an independent timeout for datachannels... since that will be confusing for JS apps (they'd see close events but the peerconnection may still be "connected")

However... while the stacks are all from association timeout, if we do this we may retransmit "forever", and in the case where we timed out, we might try to send a packet retransmit instead, and I'm concerned that this would use the bad ptr - unless the normal retransmit is blocked by the deregister/close.
(In reply to Byron Campen [:bwc] from comment #17)
> Comment on attachment 8782576 [details] [diff] [review]
> wrap and lock DataChannelConnection for SCTP callbacks
> 
> Review of attachment 8782576 [details] [diff] [review]:
> -----------------------------------------------------------------
> 
> ::: netwerk/sctp/datachannel/DataChannel.h
> @@ +296,5 @@
> >    nsCOMPtr<nsIThread> mInternalIOThread;
> > +
> > +  // Temporary wrapper object to use with sctp library callbacks so we can
> > +  // neuter it on Destroy().  Leaked on purpose when we delete this object.
> > +  ConnectionWrapper *mTempWrapper;
> 
> Oof. Leaking a mutex on purpose? Surely there's a better fix.

Leaking a mutex is better than a UAF.

However, I wonder if I can so some magic with atomics... or do some deferred cleanup of these (clean them up on a 10s timer).  Initially I'd hoped to do this with a bare ptr to minimize the leak, but that opens us to holes when trying to clear it (inherently from a different thread)
Flags: needinfo?(rjesup)
I could also use a single static Mutex for all datachannelconnections.... likely no contention (or minimal), and the leak would then just be a class ConnectionWrapper { ConnectionWrapper(foo) : mConnection(foo); RefPtr<DataChannelConnection> mConnection; }
(In reply to Randell Jesup [:jesup] from comment #18)
> Hmmm.  That will turn off detection of a failed connection.  In theory ICE
> Consent Freshness should eventually kick in and mark it as failed.  There
> might even be advantages in not having an independent timeout for
> datachannels... since that will be confusing for JS apps (they'd see close
> events but the peerconnection may still be "connected")
> 
> However... while the stacks are all from association timeout, if we do this
> we may retransmit "forever", and in the case where we timed out, we might
> try to send a packet retransmit instead, and I'm concerned that this would
> use the bad ptr - unless the normal retransmit is blocked by the
> deregister/close.

We have retransmitted multiple times before we do the cleanup. So I think
the problem is NOT related to the timerou (either HEARTBEAT or DATA retransmit),
but related to the cleanup. Which, I guess, also involves somehow the freeing
of the DTLS connection object. That is why I think my change would work around
the issue.

You can also:
* make sure you call usrscpt_close() before usrsctp_deregister().
* make sure you do not free whatever the pointer given to usrsctp points to before
  calling usrsctp_deregister().
That would (I'm guessing here) the real fix. However, it is not within the usrsctp stack.

I also think that the ICE layer detects that the peer is unreachable earlier than SCTP
does with the used default settings.
Flags: needinfo?(rjesup)
Switched to a shared StaticMutex for locking access to the DataChannelConnection pointer
Attachment #8782954 - Flags: review?(docfaraday)
Attachment #8782576 - Attachment is obsolete: true
(In reply to Michael Tüxen from comment #21)
> (In reply to Randell Jesup [:jesup] from comment #18)
> > Hmmm.  That will turn off detection of a failed connection.  In theory ICE
> > Consent Freshness should eventually kick in and mark it as failed.  There
> > might even be advantages in not having an independent timeout for
> > datachannels... since that will be confusing for JS apps (they'd see close
> > events but the peerconnection may still be "connected")
> > 
> > However... while the stacks are all from association timeout, if we do this
> > we may retransmit "forever", and in the case where we timed out, we might
> > try to send a packet retransmit instead, and I'm concerned that this would
> > use the bad ptr - unless the normal retransmit is blocked by the
> > deregister/close.
> 
> We have retransmitted multiple times before we do the cleanup. So I think
> the problem is NOT related to the timerou (either HEARTBEAT or DATA
> retransmit),
> but related to the cleanup. Which, I guess, also involves somehow the freeing
> of the DTLS connection object. That is why I think my change would work
> around
> the issue.
> 
> You can also:
> * make sure you call usrscpt_close() before usrsctp_deregister().
> * make sure you do not free whatever the pointer given to usrsctp points to
> before
>   calling usrsctp_deregister().

You mean usrsctp_deregister_address, right? Because if that will fix this, that's really easy.
Yes, I meant usrsctp_deregister_address(). If that is easy, I would say do that and let us see
if that fixes the issue in the field.
(In reply to Michael Tüxen from comment #14)
> It is for example stopped when we clear the net in the
> sctp_free_remote_addr() macro.
> 

I added an abort to this macro, and it never fires when running our mochitest suite...
|sctp_timer_stop| for |SCTP_TIMER_TYPE_SEND| is not called (transitively) by either |usrsctp_close| or |usrsctp_deregister_address|. Is there some other way this kind of timer could be getting canceled?
Flags: needinfo?(tuexen)
The retransmission timers (like all timers) are stopped in sctp_free_assoc() for each path:

	TAILQ_FOREACH(net, &asoc->nets, sctp_next) {
		(void)SCTP_OS_TIMER_STOP(&net->rxt_timer.timer);
		net->rxt_timer.self = NULL;
		(void)SCTP_OS_TIMER_STOP(&net->pmtu_timer.timer);
		net->pmtu_timer.self = NULL;
		(void)SCTP_OS_TIMER_STOP(&net->hb_timer.timer);
		net->hb_timer.self = NULL;
	}
Flags: needinfo?(tuexen)
Hmm. sctp_free_assoc() also calls:

	TAILQ_FOREACH_SAFE(net, &asoc->nets, sctp_next, nnet) {
#ifdef INVARIANTS
		if (SCTP_BASE_INFO(ipi_count_raddr) == 0) {
			panic("no net's left alloc'ed, or list points to itself");
		}
#endif
		TAILQ_REMOVE(&asoc->nets, net, sctp_next);
		sctp_free_remote_addr(net);
	}
So could it be that you don't tear down the associations in your test cases?
So, we can call close() before deregister_address() (per my comments in comment 10.)  if that's all it takes, that's easy.  
Right now, we call either close and then deregister_address, or deregister_address and then close (since they're occurring on separate threads.)  There's no way the DataChannelConnection object can be released before both have happened.  The crashes we're seeing imply that we've freed the object, and thus close and deregister_address must have completed
Flags: needinfo?(rjesup)
After calling usrsctp_deregister_address(), the usrsctp stack will still call the lower layer output routine using the provided pointer. I tested that. Avoiding that is harder than I thought, since most
of the address handling we normally use isn't used in the case where we deal with the pointer.

However, if you say that the DataChannelConnection isn't released before usrsctp_close() has been called AND usrsctp_deregister_address() has been called, then the sequence doesn't matter.

Just to be crystal clear; Here is what I get from the stack traces:

In all cases, SCTP detects via multiple retransmissions (10 to be precisely, taking several minutes of probing) that the peer is dead. Then it calls the upper layer to let him know. There you are doing some cleanups. See
https://dxr.mozilla.org/mozilla-central/source/netwerk/sctp/datachannel/DataChannel.cpp#1470
I'm not sure what happens there. Could it be that you are releasing there the DataChannelConnection?
After that callback, the SCTP stack will send out an ABORT. This is where FF crashes, since the DataChannelConnection object is gone at this point of time.

The missing piece for me is that I don't understand under which condition the DataChannelConnection gets freed (since I'm not familiar with C++)... I suspect that this might be done in the cleanup.
So that is why I'm proposing the make sure it never happens (until we understand what is going on and we have a way to test this).
So that send an ON_DISCONNECTED message to the MainThread, which calls DataChannelConnection::CloseAll().

CloseAll() calls Close() on each DataChannel, and does some cleanup on pending opens (calls Close on them).  Lastly if it closed any channels, it does SCTP_STREAM_RESET_OUTGOING (in SendOutgoingStreamReset()).

None of these should affect the DataChannelConnection object itself, which is refcounted - so long as the PeerConnection (or anything else) has a RefPtr<> to the connection, it will not be deleted.  Before the PeerConnection will drop the RefPtr (which can delete the connection), it will call Destroy() on the connection, which will at that point call close() and deregister_address().  After those have been called, it *will* destroy the connection object (we leave the sctp stack available until browser shutdown (i.e. we normally *don't* call sctp_finish() until the browser exits, though I'd like to look at doing so.)

So it looks like the original assessment is correct - there's no easy way to avoid a race between an SCTP timeout and someone destroying the connection.  Now, I don't know that's what *did* happen (i.e. that it's not some other problem), but this is a plausible if unlikely race.

So, ugly as it is, we may have to go with the original idea, since without calling usrsctp_finish there's no way to ensure that the stack won't call us (and if multiple connections in different peerconnections are active, calling usrsctp_finish would be bad.
OK. So you are saying that we have a race:
(a) One thread processing a timeout which results in terminating the SCTP association.
(b) Another thread destroying the DataChannelConnection.
Any idea how (b) can happen? Which usrsctp_*() functions might be involved?
If you have a conjecture, I could look at the code and/or write a stress tester...
Thread B is the thread that opened the SCTP library, and from previous experience (bugs) must be the one that closes it.

<user shutdown of PeerConnection>
Thread A (Mainthread): 
  Destroy() {
     Send message to Thread B to close SCTP (DestroyOnSTSThread())
     usrsctp_deregister_address()
  }
  ~DataChannelConnection() - delete of the structure (object ptr) we registered with the open

Thread B (STS):
  DestroyOnSTSThread() {
    usrsctp_close()
  }

Browser Shutdown:
  usrsctp_finish();

Because of threading, DestroyOnSTSThread() can happen before or after deregister_address.  However, you're telling me that after deregister_address "the usrsctp stack will still call the lower layer output routine using the provided pointer", so it appears there's no way to avoid that other than usrsctp_finish (which we don't do until browser shutdown and all users of sctp have shut down).

I could make the message dispatch to the STS thread synchronous (though we prefer to avoid synchronous messages starting from MainThread, since that can block the main thread of the browser) and do deregister on STS or after it returns, or I could move deregister to before the message is sent.  But from your quote above I don't think that will help me (I might still get called back using the ptr even after deregister *and* close)
Just as a side note: usrsctp_finish() will teardown the stack if there are not endpoints anymore.

Doesn't the above mean that you possibly delete the structure within thread A before thread B calls usrsctp_close().

Please note that in the crashes we are looking at, the timer thread is involved. We run the notification code within the timer thread. So the cleanup will be triggered from that thread.

Again: I'm not sure about the timeframe you need to handle this problem. If you need a quick way to avoid it, think about raising the threshold for dead peer detection.
(In reply to Michael Tüxen from comment #34)
> Just as a side note: usrsctp_finish() will teardown the stack if there are
> not endpoints anymore.
> 
> Doesn't the above mean that you possibly delete the structure within thread
> A before thread B calls usrsctp_close().

   The Runnable that calls DestroyOnSTS holds a strong ref to the DataChannelConnection.
(In reply to Michael Tüxen from comment #30)
> In all cases, SCTP detects via multiple retransmissions (10 to be precisely,
> taking several minutes of probing) that the peer is dead. Then it calls the
> upper layer to let him know. There you are doing some cleanups. See
> https://dxr.mozilla.org/mozilla-central/source/netwerk/sctp/datachannel/
> DataChannel.cpp#1470

Is there an event that signals that the SCTP stack will definitely not be calling us back anymore?
Flags: needinfo?(tuexen)
Currently not. When aborting the association, we provide the notification first and then send an ABORT.
We are currently reworking the callback API to overcome several limitations of the current way, but that needs some further testing...
Flags: needinfo?(tuexen)
Comment 35 (DestroyOnSTS holding a strong ref) means we will not destroy it until close() returns.

However - this isn't on the timer thread.  Sounds like there's a race between close/etc on other threads (close() on STS in this case) and the timer firing; i.e. some missing interlock that stops the timer from firing if close happens.

This is a security bug, and as it's a UAF we really want to take a fix for it in the next week or two so it can go out in 49 -- even if that fix is my "leak a ptr" fix
So, we have unlocked a mutex in the crashing callstack:

https://dxr.mozilla.org/mozilla-central/source/netwerk/sctp/src/netinet/sctp_callout.c#155

It does look like this leaves a brief window that could allow the Destroy() code to finish at the same time the timer is popping.
Is there some way we could get the sctp stack to schedule an immediate timer that comes around as a callback on the timer thread? We could schedule this timer in DestroyOnSTS (and move the deregister there too), and once that timer pops we can dispose of the DataChannelConnection.

If there is not currently a way to do this, how hard would it be to make one?
Flags: needinfo?(tuexen)
(In reply to Byron Campen [:bwc] from comment #39)
> So, we have unlocked a mutex in the crashing callstack:
> 
> https://dxr.mozilla.org/mozilla-central/source/netwerk/sctp/src/netinet/
> sctp_callout.c#155
> 
> It does look like this leaves a brief window that could allow the Destroy()
> code to finish at the same time the timer is popping.
The TIMERQ_LOCK just protects the list of timers. If the stored timer is stopped,
sctp_os_timer_next is adopted.
Flags: needinfo?(tuexen)
(In reply to Byron Campen [:bwc] from comment #40)
> Is there some way we could get the sctp stack to schedule an immediate timer
> that comes around as a callback on the timer thread? We could schedule this
> timer in DestroyOnSTS (and move the deregister there too), and once that
> timer pops we can dispose of the DataChannelConnection.
> 
> If there is not currently a way to do this, how hard would it be to make one?

It is just software, so it can be changed.
However, the code has evolved and the base you are using is over a year old.
I would have to look at that code and adding such a timer without having a reproducible
test case to make sure we are working around that issue is something I would like
to avoid. I guess the simplest workaround for the crashes I have seen is to disable the
dead peer detection at the SCTP layer for now. Is there a reason not going down this path?
If Randell feels save to add his patch, it can be added in addition to setting the maximum
number of retransmits to 0xffffffff.
My only concern with the "retransmit forever" case is this:
> However... while the stacks are all from association timeout, if we do this we may retransmit
> "forever", and in the case where we timed out, we might try to send a packet retransmit instead, and
> I'm concerned that this would use the bad ptr - unless the normal retransmit is blocked by the
> deregister/close.

If close/deregister_address will block *that* timer (the retransmit) - then it could help.
All stack traces I have looked at showed that sctp_abort_an_association() was called from sctp_threshold_management. So it was always happening when SCTP was destroying the association,
not when it did all the retransmits before.
That is why I belief that it is related to the cleanup when handling dead peer detection.
I have not seen a trace where a normal retransmission triggered the crash. Have you?
OK. By looking at the code again, this might be a fix for the issue:

Move the "notify the ulp" code in
https://dxr.mozilla.org/mozilla-central/source/netwerk/sctp/src/netinet/sctputil.c#4207
below the "notify the peer" code in
https://dxr.mozilla.org/mozilla-central/source/netwerk/sctp/src/netinet/sctputil.c#4211

This would mean we send out the ABORT (using your object pointer) before notifying
the C++ code, which might free the pointer.

For consistency, We would need a similar swap around
https://dxr.mozilla.org/mozilla-central/source/netwerk/sctp/src/netinet/sctputil.c#4066

However, without a test, we can't be sure. But it might be worth a try... At least
I would try to test it. What do you think?
Flags: needinfo?(rjesup)
It makes sense in any case, since when you notify the owner of the association failure it may break the channel between usrsctp and the network.  (this becomes an issue once sctp travels in a tunnel via the callbacks, as we use it).  As to if this could cause the problem... maybe.  We will fire close events to JS, and it may well tear down the peerconnection in response.
Flags: needinfo?(rjesup)
I think it's worth landing this in any case.  I forget the details of the analysis, but perhaps this causes the race with a JS app that responds to onclose by tearing down the peerconnection.
Attachment #8783983 - Flags: review?(docfaraday)
Comment on attachment 8783983 [details] [diff] [review]
Swap order of notifications on association failure

Review of attachment 8783983 [details] [diff] [review]:
-----------------------------------------------------------------

This might do the trick? Worth a shot at any rate.
Attachment #8783983 - Flags: review?(docfaraday) → review+
Attachment #8782954 - Flags: review?(docfaraday)
Comment on attachment 8783983 [details] [diff] [review]
Swap order of notifications on association failure

[Security approval request comment]
How easily could an exploit be constructed based on the patch? Very hard

Do comments in the patch, the check-in comment, or tests included in the patch paint a bulls-eye on the security problem? No - at most it is only of interest if it lands on a sec bug and is uplifted.  Also, patch has landed upstream - we could land it under a generic "cherry-pick mod from upstream" bug if we care.

Which older supported branches are affected by this flaw? All, though it appears as if other changes recently exposed the flaw (timing/references/etc).

If not all supported branches, which bug introduced the flaw? N/A

Do you have backports for the affected branches? If not, how different, hard to create, and risky will they be? Trivial

How likely is this patch to cause regressions; how much testing does it need? Very low chance of regression; swaps the order of notifications.
Attachment #8783983 - Flags: sec-approval?
Attachment #8783983 - Flags: approval-mozilla-beta?
Attachment #8783983 - Flags: approval-mozilla-aurora?
Likely affects 48 and ESR45, but they may not have changes elsewhere that expose the issue.  They might still be exploitable though.
Comment on attachment 8783983 [details] [diff] [review]
Swap order of notifications on association failure

Approvals given. This is a tiny patch. We should take this on ESR45 as well.
Attachment #8783983 - Flags: sec-approval?
Attachment #8783983 - Flags: sec-approval+
Attachment #8783983 - Flags: approval-mozilla-beta?
Attachment #8783983 - Flags: approval-mozilla-beta+
Attachment #8783983 - Flags: approval-mozilla-aurora?
Attachment #8783983 - Flags: approval-mozilla-aurora+
Comment on attachment 8783983 [details] [diff] [review]
Swap order of notifications on association failure

[Approval Request Comment]
User impact if declined: sec-crit fixed on other branches

Fix Landed on Version: 51 (uplift to 49)

Risk to taking this patch (and alternatives if risky): very low risk

String or UUID changes made by this patch: none

See https://wiki.mozilla.org/Release_Management/ESR_Landing_Process for more info.
Attachment #8783983 - Flags: approval-mozilla-esr45?
Note: We might hold landing it on ESR45 until we see some crashstats data that confirms the fix in the field.
https://hg.mozilla.org/mozilla-central/rev/fab696bc92fd
Status: NEW → RESOLVED
Closed: 8 years ago
Flags: in-testsuite?
Resolution: --- → FIXED
Target Milestone: --- → mozilla51
sec critical, taking it to esr 45. Should be in 45.4
Attachment #8783983 - Flags: approval-mozilla-esr45? → approval-mozilla-esr45+
Group: media-core-security → core-security-release
Whiteboard: [adv-main49+][adv-esr45.4+]
7 crashes with this signature in Socorro n the last week, the most recent affected build being 49.0b6. 
https://crash-stats.mozilla.com/signature/?product=Firefox&signature=mozilla%3A%3ADataChannelConnection%3A%3ASctpDtlsOutput

Closing this issue.
Blocks: 1311380
Group: core-security-release
You need to log in before you can comment on or make changes to this bug.

Attachment

General

Created:
Updated:
Size: