openzipkin / zipkin

Zipkin is a distributed tracing system
https://zipkin.io/
Apache License 2.0
17.01k stars 3.09k forks source link

Support async spans #1243

Closed codefromthecrypt closed 7 years ago

codefromthecrypt commented 8 years ago

We've had many issues about representing one-way activity. This is an attempt to define the most semantic means to address this concern in zipkin v1.

In zipkin, we have an RPC span, where two hosts participate. They participate by adding bookend annotations to a single span. This allows for many conveniences including hosts able to report incomplete work, and report in any order.

For example, a typical RPC span is ["cs", "sr", "ss", "cr"], for a dual host (client+server) span, or either of its partial ones ["sr", "ss"](uninstrumented client) and ["cs", "cr"](uninstrumented server).

The path of least resistance for async (half-duplex, or messaging), is to follow this model. This prevents us from needing to wait for a big-bang zipkin 2 release of the entire ecosystem.

My proposal is to add a messaging span: "ms" and "mr".

The message sender adds a "ms" when a one-way message is sent. They are free to add "ws" to indicate it going on the wire.

The message receiver adds a "mr" when a one-way message is received. They are free to precede that with "wr" to indicate one being processed from the wire.

This works like the RPC spans, but to pin it down, here are example spans.

["mr"] <- message received from an uninstrumented sender. The host of "mr" is the processor. ["ms"] <- message sent to an uninstrumented recipient, or the message was dropped. The host of "mr" is the sender. ["ms", "mr"] <- message sent and received from instrumented hosts. Similarly, the host of "ms" is the sender and the host of "mr" is the receiver.

The duration of a complete messaging span can be a lot longer than an RPC span, potentially orders of magnitude longer than the processing time. Once we have a means to indicate this (via "ms" "mr"), we can teach the UI to collapse the timeline or do other tricks.

Messaging spans need to be propagated, just like RPC ones. For example, storing the binary version of a trace id as metadata on the message is appropriate.

This design adds pressure for "linked traces" especially when messages are needed to create a new outbound activity. This is out of scope of this design, but not precluded by it. In other words, we need to take on "linked traces" separately.

Fixes #1189 Fixes #925

codefromthecrypt commented 8 years ago

related issue: multiple parents aka linked traces https://github.com/openzipkin/zipkin/issues/1244

codefromthecrypt commented 8 years ago

had a chat with @mbogoevici from spring-cloud-stream.

We talked about a special case, where you might want to visualize the trace as being "open" the same way that a root span normally works in RPC. IE the critical path of the pipeline continues until a terminal event occurs. This allows you to see the entire duration clearly.

This implies the root span isn't closed immediately when it sent something to a broker, rather it is left open until a terminal event occurs. Something down the line will close it.

  1. Report a start annotation in the root span like "pipeline.begin" or similar
  2. When going across a broker, propagate the root span id as baggage and report "ms" in the current span.
  3. When receiving from a broker, extract the current span and report "mr"; extract the root span id from baggage and propagate in process
  4. repeat 2, 3 as necessary
  5. When you've hit a terminal point for the pipeline, report "pipeline.end" or similar using the root span id you propagated

cc @dsyer @marcingrzejszczak

lijunyong commented 7 years ago

@adriancole when publish zipkin 2 release,it support MQ and asyn call

marcingrzejszczak commented 7 years ago

Looks nice. So whenever I use messaging we should go with this approach or is only for fully flow-like applications? Meaning that if you have a simple queue based communication between two apps should we use the same approach?

sslavic commented 7 years ago

Would be nice if both push (notification/topic) and pull (queue) communication was covered.

One tricky thing, which would also be nice to be covered as well, is batching support in both produce (publish), and consume side (push, pull). E.g. one can have multiple publish requests which end up consumed in one pull request. Also, single publish request could publish multiple messages, which get delivered with multiple independent push requests.

codefromthecrypt commented 7 years ago

@adriancole https://github.com/adriancole when publish zipkin 2 release

we usually discuss when features will land vs certain version numbers. What features are you interested in?

codefromthecrypt commented 7 years ago

Would be nice if both push (notification/topic) and pull (queue) communication was covered.

One tricky thing, which would also be nice to be covered as well, is batching support in both produce (publish), and consume side (push, pull). E.g. one can have multiple publish requests which end up consumed in one pull request. Also, single publish request could publish multiple messages, which get delivered with multiple independent push requests.

good point. we've actually another issue on that topic, as it is trickier than one-way single producer single consumer: linked traces https://github.com/openzipkin/zipkin/issues/1244

lijunyong commented 7 years ago

@adriancole I interested trace MQ and asyn call

lijunyong commented 7 years ago

@adriancole The path of least resistance for async (half-duplex, or messaging), is to follow this model. This prevents us from needing to wait for a big-bang zipkin 2 release of the entire ecosystem.

mbogoevici commented 7 years ago

@marcingrzejszczak @adriancole the initial discussion was in the context of Stream apps - we like to have a continuous trace that covers the entire pipeline - but I guess most of the reasoning applies in the general case of apps talking to each other over a broker - including linked spans for pub-sub with multiple consumers

mbogoevici commented 7 years ago

In terms of an update, I have a working POC for Sleuth sans Baggage which I will happy to share once I change it to use the Baggage support that @marcingrzejszczak has added in 1.2

codefromthecrypt commented 7 years ago

@adriancole https://github.com/adriancole The path of least resistance for async (half-duplex, or messaging), is to follow this model. This prevents us from needing to wait for a big-bang zipkin 2 release of the entire ecosystem.

gotcha.. no big bangs here (though there are some features that take a while!) :)

the effort to add messaging span is very little. It is defining constants and adjusting some things in the UI and clock skew checker. Main reason I delayed this one was rule-of-three. I think we have enough interested at this point to move forward with implementing "ms" "mr"

lijunyong commented 7 years ago

@adriancole What‘s the plans,probably when released,Thanks a lot

codefromthecrypt commented 7 years ago

I'll update this with a pull request in the next couple days latest by monday. In the mean time, you can always use the "ms" and "mr" annotations yourself, they just won't be interpreted by the UI until we upgrade.

codefromthecrypt commented 7 years ago

@lijunyong ps after thinking about it.. I am not going to raise this PR yet. @mbogoevici's work is new information. We should see if that is extending this or an alternative. Plus I'd like to get a tally of at least a couple folks who aren't primarily interested in JVM environments. Finally, we need to vet this against single-host span pattern.

If there's anyone on this thread tracing non java environments with opinions on one-way tracing please speak up (cc @jcarres-mdsol @ys @mjbryant @hammett)!

Between now and when pull requests to various repos are reviewed and merged, feel free to just use the conventions described for one-way messages. Ex. send "ms" on the way out and "mr" when receiving (just like client+server, except no response is carried back). There's nothing in zipkin that prevents experimenting with this.

hammett commented 7 years ago

Message based (one way) is 95% of our system. I'd be very interested in native support for this.

lijunyong commented 7 years ago

@adriancole I don't understand,Could you provide the example about MQ in github?Thanks a lot

jcarres-mdsol commented 7 years ago

So this will not mix with the current system? For instance: User -> System 1(trace1, sr cs,cr, ss) -> System2(trace1 sr ms, ss) -----> System3(trace1 mr cs,cr) -> System3(sr ss)

Instead you are saying that ms involves creating a trace2?

codefromthecrypt commented 7 years ago

@lijunyong @jcarres-mdsol here's a repaste of a comment on brave which I hope will clarify both of your questions. Please let me know if it doesn't.

Here's an example scenario asked by @coolbetm

Consider the following simplish example: Suppose service A receives a post to its REST endpoint and produces the body of the post to a kafka topic. Service B is tailing on that topic and consumes it, parses it, and produces a derived message to second topic. Service C is tailing on the second topic, receives the derived message, and responds by executing a database update. What do you imagine the zipkin trace should look like (what spans are there and in what order do they start/end)?

Service A receives a POST and sends a message to kafka {traceid=1, parent=_, spanid=1 [sr=serviceA, ss=serviceA]} // server span {traceid=1, parent=1, spanid=2 ["make message", lc=serviceA]} // local span {traceid=1, parent=2, spanid=3 [ms=serviceA]} // messaging span sent

Service B consumes the message and puts a message on a second topic {traceid=1, parent=2, spanid=3 [mr=serviceB]} // messaging span received (same span id) {traceid=1, parent=3, spanid=4 ["process and make another message", lc=serviceB]} // local span {traceid=1, parent=4, spanid=5 [ms=serviceB]} // messaging span sent

Service C consumes the message and saves to a database {traceid=1, parent=4, spanid=5 [mr=serviceC]} // messaging span received (same span id) {traceid=1, parent=5, spanid=6 ["process message", lc=serviceC]} // local span {traceid=1, parent=6, spanid=7 [cs=serviceC, cr=serviceC, sa=database]} // client span

jcarres-mdsol commented 7 years ago

Why the last CS/CR to itself?

codefromthecrypt commented 7 years ago

Why the last CS/CR to itself?

This is a call to a database. I'm assuming the database doesn't speak zipkin, so we just log "sa" in the client span to indicate the remote side.

jcarres-mdsol commented 7 years ago

If serviceC finishes processing the message but does not generate any other zipkin information after mr (for instance we never trace calls to DBs) how will we know the trace has in fact ended?

codefromthecrypt commented 7 years ago

If serviceC finish processing the message but does not generate any other zipkin information after mr (for instance we never trace calls to DBs) how will we know the trace has in fact ended?

Since "mr" closes the messaging span, all spans are closed.. so it is the same situation as today. A trace is effectively ended when all spans finish.

jcarres-mdsol commented 7 years ago

I thought mr was emitted when the worker got the message from the queue. Performance wise I would want to know two things:

To do both we need at least one more annotation, no?

codefromthecrypt commented 7 years ago

I thought mr was emitted when the worker got the message from the queue. Performance wise I would want to know two things:

  • How long the message was in the messaging bus

After "ms" we can use the normal "ws" to indicate when it went to the bus. Similarly, we can use "wr" to tell the difference between when we got the message vs started to process it.

  • How long it took the worker to process the message

The messaging span in my mind really should be transit only (don't record message processing in the span). That's why in my example, I have listed a child span to process the message. Timewise, this directly follows the messaging span, so its (start) timestamp will very close or exactly the same as "mr" of its parent (the messaging span)

To do both we need to tell at least one more annotation, no?

yeah I think probably I didn't highlight "ws" "wr" (which incidentally are defined also for client and server spans, just not widely used)

jcarres-mdsol commented 7 years ago

Ah, ok, then ServiceC would also use lc to specify how long its processing took

codefromthecrypt commented 7 years ago

Ah, ok, then ServiceC would also use lc to specify how long its processing took

heh.. whoops. guess I left that out of the span list. fixing!

codefromthecrypt commented 7 years ago

ok updated https://github.com/openzipkin/zipkin/issues/1243#issuecomment-266359841 .. I understand why it was unclear now! thanks @jcarres-mdsol

lijunyong commented 7 years ago

@adriancole should I add span(traceid, parent, spanid) in message header,then send the json data[add "ms" and "mr" in annotations] to zipkin by spanCollector(HttpSpanCollector),but the duration how to calculate like rest by interceptor? image

codefromthecrypt commented 7 years ago

@adriancole https://github.com/adriancole should I add span(traceid, parent, spanid) in message header,then send the json data to zipkin by spanCollector(HttpSpanCollector)?

Yes, you can use the same propagation and reporting approach as client/server, just where you encode the trace identifiers will be messaging specific. Some are using message headers exactly as they would if they were http (spring-cloud-sleuth does this, for example).

mbogoevici commented 7 years ago

@adriancole re https://github.com/openzipkin/zipkin/issues/1243#issuecomment-266359841

This is largely consistent with what we envisioned for Spring Cloud Stream - with the addition of propagating the root id of the sequence as baggage and sending additional spans towards it as messages enter input channels and exit output channels, so that we can have a span that covers the entire duration of a logical stream.

I like the differentiation between ws/wr ms/mr as we might use that to capture the time spent with pre-send/post-receive activities (e.g. unmarshalling, etc) as opposed to raw time spent in the broker. In my POC the root span remained parent, but I guess I'll have to update that as the lineage is more correct by setting parents sequentially.

The other thing I can think of is identifying the message system as a separate endpoint - i.e. using the broker as the endpoint (e.g. 'kafka' or something else - for Stream we can adopt the logical binder name, in other cases it could be the broker name) and the logical destination as the span name, e.g.

{traceid=1, parent=, spanid=1 [sr=serviceA, ss=serviceA]} // server span {traceid=1, parent=1, spanid=2 ["make message", lc=serviceA]} // local span {traceid=1, parent=2, spanid=3 [ms=kafka]} // messaging span sent_

Service B consumes the message and puts a message on a second topic {traceid=1, parent=2, spanid=3 [mr=kafka]} // messaging span received (same span id) {traceid=1, parent=3, spanid=4 ["process and make another message", lc=serviceB]} // local span {traceid=1, parent=4, spanid=5 [ms=kafka]} // messaging span sent

Service C consumes the message and saves to a database {traceid=1, parent=4, spanid=5 [mr=kafka]} // messaging span received (same span id) {traceid=1, parent=5, spanid=6 ["process message", lc=serviceC]} // local span {traceid=1, parent=6, spanid=7 [cs=serviceC, cr=serviceC, sa=database]} // client span

Then for instance you could search for spans involving 'kafka'. WDYT?

codefromthecrypt commented 7 years ago

The other thing I can think of is identifying the message system as a separate endpoint - i.e. using the broker as the endpoint (e.g. 'kafka' or something else - for Stream we can adopt the logical binder name, in other cases it could be the broker name) and the logical destination as the span name, e.g.

{traceid=1, parent=_, spanid=1 [sr=serviceA, ss=serviceA]} // server span {traceid=1, parent=1, spanid=2 ["make message", lc=serviceA]} // local span {traceid=1, parent=2, spanid=3 [ms=kafka]} // messaging span sent

Service B consumes the message and puts a message on a second topic {traceid=1, parent=2, spanid=3 [mr=kafka]} // messaging span received (same span id) {traceid=1, parent=3, spanid=4 ["process and make another message", lc=serviceB]} // local span {traceid=1, parent=4, spanid=5 [ms=kafka]} // messaging span sent

Service C consumes the message and saves to a database {traceid=1, parent=4, spanid=5 [mr=kafka]} // messaging span received (same span id) {traceid=1, parent=5, spanid=6 ["process message", lc=serviceC]} // local span {traceid=1, parent=6, spanid=7 [cs=serviceC, cr=serviceC, sa=database]} // client span

Then for instance you could search for spans involving 'kafka'. WDYT?

Agreed on the concern, but we still need to set mr's endpoint to the local service receiving from kafka.

One way is to add "ma" which is similar to client address and server address, except it represents the broker (using m for congruence with the timestamp annotations). In zipkin v2 it would simply be the remote address (no extra type info needed)

mbogoevici commented 7 years ago

we still need to set mr's endpoint to the local service receiving from kafka.

OK, I see the weak point of my proposal :). I suppose 'ms' then gets set to the sender, then ?

One way is to add "ma" which is similar to client address and server address, except it represents the broker (using m for congruence with the timestamp annotations).

Great - I think it would be nice to emphasize that those spans are traversing the broker (e.g. Kafka) as opposed to being either a sending or a receiving service span.

codefromthecrypt commented 7 years ago

I suppose 'ms' then gets set to the sender, then ?

yep. all timestamp annotations are attributed to the host that caused the activity

Great - I think it would be nice to emphasize that those spans are traversing the broker (e.g. Kafka) as opposed to belonging to the sending and receiving service.

agreed!

codefromthecrypt commented 7 years ago

first step: adding constants to the thrift https://github.com/openzipkin/zipkin-api/pull/24

codefromthecrypt commented 7 years ago

During https://github.com/openzipkin/zipkin-api/pull/24 we noticed this can be complicated when the messaging system is also RPC (such as pubsub or Amazon SQS which are both http services). For example, how would they know if they should add a "ms" or "cs" considering the production of a message is literally an rpc (usually http) call? I noted that we still have the option to not do this change, or to instead repurpose "cs" "sr" to address one-way concerns. For example, thrift and other RPC systems have a one-way specialization.

It is important to think about things like this, but also understand tactical vs strategic change. For example, any annotations like this are tactical in my view. I'll try to recap some of this (which came from a gitter chat with @jcarres-mdsol)


I think that linked traces are a better fit for messaging than these annotations. The issue with linked traces is we need to change schema (which isn't a problem, just something to get through)

The "ms" "mr" was a way to identify a means to handle one-way for those in the brown field, because it requires very little change (similar to if we decided to treat "cs" "sr" missing response, as a special one-way). So, presence of some heuristic for one-way doesn't preclude linked traces, or even handle it better in any way. It is merely a stepping stone.

Here are the steps I'm thinking of that lead us forward with regards to messaging:

Both of the above are independent, but once the latter is in, we can add linked traces to it much easier than trying to add to the multi-host span model we have today.

Back to the effort here, as far as I can tell, we either proceed as noted or try to repurpose the "cs" and "sr" similarly. Another option is to not do this and head directly towards links.

I'll give another couple days to make sure people have seen this and can weigh in if any of this context changes their mind.

codefromthecrypt commented 7 years ago

I think we should make a decision as perpetually incomplete work is haunting, particularly as we've discussed this topic for a very long time now, and there are people interested in it. That, or we should make visible what must happen before a decision can be made. We don't want a worst of show, where we've spent lots of time discussing, and gathering interest, but lose them in analysis paralysis.

A valid question is what to do when an RPC system is used to carry what is semantically a one-way message. I'd flatten this to the generic topic of one-way RPC, which exists in many systems (regardless if you pay attention to acks or not).

This is separate from the concern that led to this issue, which is that a non-RPC system is used to carry single-producer single-consumer messages. The latter concern is addressed by ad-hoc users today, using annotations which may or may not be named "ms" and "mr" :)

If we feel those in the latter camp should continue to work ad-hoc, that's a valid thing, but I think people should be more explicit about the desire to halt this, and what, if anything it would be replaced with. We'd then close the issue as won't fix and stop spending time on it.

If we feel there's something that must be addressed, let's make test cases for that. They can be simple descriptions. Regardless, we need to find a way to progress this topic.

codefromthecrypt commented 7 years ago

Here's a diagram of the latest revision of this, based on a pre new-year chat with @jcarres-mdsol

By not defining a new address annotation for the message broker, a lot of code will be similar to today. Should we decide to use new annotations, they would be exactly like existing, except clarifying one-way semantics.

that reduces the explanation to... "ms" is like "cs" except there's no response anticipated "mr" is like "sr" except we aren't sending a response

   Producer Tracer                                    Consumer Tracer     
+------------------+                               +------------------+
| +--------------+ |     +-----------------+       | +--------------+ |
| | TraceContext |======>| Message Headers |========>| TraceContext | |
| +--------------+ |     +-----------------+       | +--------------+ |
+--------||--------+                               +--------||--------+
   start ||                                                 ||
         \/                                          finish ||
span(context).annotate("ms")                                \/
             .address("sa", broker)          span(context).annotate("mr")
                                                          .address("ca", broker)
codefromthecrypt commented 7 years ago

NEXT STEP: Experiment with this design both in server and instrumentation projects

One danger area I think there is will be the code involved to handle "ms", "mr" neatly vs just allowing "cs" and "sr". The code impact of this will be present in instrumentation as well as the server (and its UI). For example, instrumentation will amplify this work, as a dozen different tracers may need new if statements etc.

We need to compare implementing one-way/unary/messaging as..

In this repo, we can open a branch/pull request would include test cases that include one-way spans, modeling pubsub with pubsub RPC on either side, checking the dependency linker and the clock skew adjuster. This pull request should note experiences that led to either option being preferable.

In one or more instrumentation repos, we'd check the two modeling approaches (cs/sr with flush vs ms/mr implicit flush), and in their corresponding pull request note experiences.

Once we have the bigger picture, I think we can lock-in and roll out what works best. I'm glad the choices are quite narrow at this point.

@openzipkin/instrumentation-owners anyone have time to help experiment on this?

codefromthecrypt commented 7 years ago

In brave 4, I think it is easier overall to use "cs" and "sr". I made an example with Kafka under this assumption https://github.com/openzipkin/brave/pull/300/files#diff-0f684bf482d2b5acd65765f8d6f8f24eR35

codefromthecrypt commented 7 years ago

moved the kafka example to a gist as I think it is too heavyweight to be in the normal test tree https://gist.github.com/adriancole/76d94054b77e3be338bd75424ca8ba30

codefromthecrypt commented 7 years ago

I think it is much better to formalize one-way/async as client+server without a response. It is feasible and it is more correct to say one-way/async than say "messaging" knowing we only handle single-producer single-consumer anyway. The linked spans work could pick up from there.

I've added a demo in Brave: https://github.com/openzipkin/brave/tree/master/brave/src/test/java/brave/features/async And I've also verified the dependency graph pretty-much works as-is: https://github.com/openzipkin/zipkin/pull/1478

I know the clock skew adjuster will change, but it will be less work than adding new annotations, as it already special cases skew found when "sr" happens before "cs".

I thought that it might break single-host spans in the dependency graph until I realized that single-host spans are already broken! In other words, regardless of one or two-way, the dependency graph would need code anyway.

So summary is that we should forget about "ms"->"mr" and go with "cs"->"sr", which reduces work on all parties to support this feature. If anyone feels against this, please make it known!

codefromthecrypt commented 7 years ago

ps I chatted with @fedj and don't think we can help correct clock skew on async-spans. The reason is that the current approach uses the two views of duration in the same span to interpret skew. Since async would have only a single view of latency, it won't work. This means either people need to keep their clocks in sync, or they need to resolve skew as a part of the reporting pipeline somehow.

The problem will be the same regardless of which annotations are used except in the case where messaging is implemented under the covers as normal RPC. In that case, all 4 annotations will be present and then skew can correct automatically.

codefromthecrypt commented 7 years ago

Another topic of interest when we get to documenting. If someone is trying to model an "ack", they should be careful not to use "cr" since it isn't semantically the same (as cr implies a round trip of an operation). When modeling ack, it would be better to use "ws" wire send or just make up another annotation (ex. kafka.ack).

codefromthecrypt commented 7 years ago

added #1497 to make one-way spans via cs -> sr work

codefromthecrypt commented 7 years ago

docs https://github.com/openzipkin/openzipkin.github.io/pull/76

codefromthecrypt commented 7 years ago

fyi, those who need to trace kafka via one-way approach can likely do so using 0.11 Brave implementation may follow here https://github.com/openzipkin/brave/issues/211

codefromthecrypt commented 7 years ago

In a recent tracing workshop, we were able to elaborate a bit on async tracing. In many cases, new traces are not needed. Ex if propagation does not imply sharing spans, you can do single producer multiple consumer as normal spans in the same trace. Here's a diagram describing most concerns

    ┌──────────────────────────────┐             ┌───────────────────────────────────────────┐
    │   RPC with Async Callbacks   │             │                 Messaging                 │
    ├──────────────────────────────┤             ├───────────────────────────────────────────┤
    │                              │             │                                           │
    │ request-response             │      ┌──────▶ Single Producer Single Consumer           │
    │                              │      │      │                                           │
    │   deferring completion       │      │      │ No guarantee of only one consumer from    │
    │    onto a different thread   │      │      │ the POV of the producer. No guarantee     │
    │                              │      │      │ of the scale of latency between producer  │
    │                              │      │      │ and consumer. Can consider collapsing     │
    │                              │      │      │ part of the timeline.                     │
    │ one-way                      │      │      │                                           │
    │                              │      │      │                                           │
    │   breaks assumptions of      │      │      │ Single Producer Multiple Consumer         │
    │   caller knowing when the    │──────▶      │                                           │
    │   request finished           │             │ Breaks shared span model, ex each consumer│
    │                              │             │ may accidentally use the same span ID     │
    │   depending on POV knowing   │             │ Harder to visualize if there are varied   │
    │   a request finished may     │             │ time scales per consumer. That said, each │
    │   be less important than if  │             │ consumer can still work as a child span   │
    │   a caller was blocked       │             │                                           │
    │                              │             │                                           │
    │                              │             │                                           │
    └──────────────────────────────┘             └───────────────────────────────────────────┘

    ┌─────────────────────────────────┐           ┌───────────────────────────────────────────┐
    │            Batching             │           │        Distributed Flow Processing        │
    ├─────────────────────────────────┤           ├───────────────────────────────────────────┤
    │                                 │           │                                           │
    │ Examples are coallesced writes  │           │ Inter-process Streams (like Spark)        │
    │ or flushing. A batch can be a   │           │                                           │
    │ separate trace with links added │           │ The user can define a join/zip/etc which  │
    │ for each enclosed element. The  │           │ may combine work started in multiple      │
    │ root is the batch. Each inserted│           │ traces. Phases tend to be coarse grained  │
    │ span is a direct child.         │           │ which maps to fewer spans. Challenge is   │
    │                                 │           │ if we can detect or assist in how to      │
    │ Representing the batch explains │           │ represent this (ex like we do in batching)│
    │ worst case latency like if there│           │                                           │
    │ was a long write ahead of you   │           │ How do we address latency of a data flow? │
    │                                 │           │                                           │
    │ Also could use correlation id   │           └──────────────────┬────────────────────────┘
    │ for the batch or another parent │                              │
    │                                 │                              │
    │                                 │                              │
    │                                 │                              │
    │                                 │                              │
    └─────────────────────────────────┘                              │
                                                                     │
                                                        ┌────────────┴───────────────────────┐
┌───────────────────────────────────────────┐           │                                    │
│   Intra-process Streams (like Reactor)    │           │Apis look the same, but challenges  │
├───────────────────────────────────────────┤           │can be different. For example,      │
│                                           │           │reactive streams define threading   │
│ Intra-process Streams (like Reactor)      │           │semantics which can be helpful in   │
│ Automatic instrumentation challenges:     │           │understanding how an application    │
│ Granularity challenges include small as   │     ┌─────┤behaves. Data flows have coarser    │
│ clock ticks. Infinite sequences exist.    ├─────┘     │granularity which maps easier to    │
│ No way to tell by type if a flux will end │           │separate spans.                     │
│ and many operators dont have metadata to  │           │                                    │
│ indicate if it is a grouping command or   │           │Both break models or visualizations │
│ not. Notable exception is parallel flux.  │           │that assume a tree shape (ex joins  │
│                                           │           │are possible in either). Both can be│
│ In cases where you create the source, ex  │           │infinite, though each process w/in  │
│ web flux, you can know for sure it is     │           │the flow has a beginning and end.   │
│ finite and how model spans on behalf of   │           │                                    │
│ requests.                                 │           └────────────────────────────────────┘
│                                           │
│ Manual instrumentation is possible where  │
│ you can materialize a context which will  │
│ be available at any point in the flux.    │
└───────────────────────────────────────────┘
codefromthecrypt commented 7 years ago

closing as async one-way is done. lets move to https://github.com/openzipkin/zipkin/issues/1654 for multiple consumers