Open Sgitario opened 6 months ago
/cc @alesj (kafka), @cescoffier (kafka), @ozangunalp (kafka)
To give more context: https://github.com/smallrye/smallrye-reactive-messaging/discussions/2609#discussioncomment-9366838 I haven't checked the reproducer but it is most likely a limitation on Smallrye reactive messaging upstream.
To give more context: smallrye/smallrye-reactive-messaging#2609 (reply in thread) I haven't checked the reproducer but it is most likely a limitation on Smallrye reactive messaging upstream.
If it can't be fixed/supported, I think this should be noted in the documentation as an incompatibility between these two extensions. I would say this is a major limitation that most users tend to oversee/expect to be working.
Hey @Sgitario,
I am currently experimenting with a very similar usecase. I want to try to get away with a outbox pattern light as described here: https://spring.io/blog/2023/10/24/a-use-case-for-transactions-adapting-to-transactional-outbox-pattern.
Here's what I've come up so far based on your example - Approach 3 - and this hint: https://smallrye.io/smallrye-mutiny/latest/guides/imperative-to-reactive/#running-blocking-code-on-subscription. Note that in my personal working example instead of hibernate I am using jdbc only, so not sure if this approach works with hibernate.
I would love to hear your opinion on this though :). I am especially unsure about the internalProcessing.flatMap(...) call as I have no experience with using Mutiny so far.
@ApplicationScoped
public class NewAuthorConsumer {
@Channel("new-book-out")
KafkaTransactions<String> kafkaTx;
@Incoming("new-author-in")
Uni<Void> consume(Message<String> message) {
Log.infof("Received new message in authors topic with name `{}`", message.getPayload());
// what we do in our application wrapped in a boolean
// that indicates if reply message should be sent
Uni<Boolean> internalProcessing = Uni.createFrom()
// wrapped up in a boolean indicating successful creation
.item(processNewAuthor(message.getPayload()))
// when author is already processed
// assume reply message has already been sent; therefore ...
.onFailure(this::isAlreadyProcessed).recoverWithItem(false)
// this is the important part which will delegate to a worker thread
.runSubscriptionOn(Infrastructure.getDefaultWorkerPool());
return kafkaTx.withTransactionAndAck(message, txEmitter ->
internalProcessing.flatMap(authorAdded -> {
if (authorAdded) {
txEmitter.send(message.getPayload() + "-book1");
}
return Uni.createFrom().voidItem();
}));
}
public boolean processNewAuthor(String authorName) {
persistAuthor(authorName);
return true;
}
@Transactional
public void persistAuthor(String name) {
Author author = new Author();
author.name = name;
author.persist();
}
private boolean isAlreadyProcessed(Throwable t) {
return t instanceof IntegrityConstraintViolationException
&& t.getCause() != null
&& t.getCause().getMessage().contains("author_name_key");
}
}
Hey @Sgitario,
I am currently experimenting with a very similar usecase. I want to try to get away with a outbox pattern light as described here: https://spring.io/blog/2023/10/24/a-use-case-for-transactions-adapting-to-transactional-outbox-pattern.
Here's what I've come up so far based on your example - Approach 3 - and this hint: https://smallrye.io/smallrye-mutiny/latest/guides/imperative-to-reactive/#running-blocking-code-on-subscription. Note that in my personal working example instead of hibernate I am using jdbc only, so not sure if this approach works with hibernate.
I would love to hear your opinion on this though :). I am especially unsure about the internalProcessing.flatMap(...) call as I have no experience with using Mutiny so far.
@ApplicationScoped public class NewAuthorConsumer { @Channel("new-book-out") KafkaTransactions<String> kafkaTx; @Incoming("new-author-in") Uni<Void> consume(Message<String> message) { Log.infof("Received new message in authors topic with name `{}`", message.getPayload()); // what we do in our application wrapped in a boolean // that indicates if reply message should be sent Uni<Boolean> internalProcessing = Uni.createFrom() // wrapped up in a boolean indicating successful creation .item(processNewAuthor(message.getPayload())) // when author is already processed // assume reply message has already been sent; therefore ... .onFailure(this::isAlreadyProcessed).recoverWithItem(false) // this is the important part which will delegate to a worker thread .runSubscriptionOn(Infrastructure.getDefaultWorkerPool()); return kafkaTx.withTransactionAndAck(message, txEmitter -> internalProcessing.flatMap(authorAdded -> { if (authorAdded) { txEmitter.send(message.getPayload() + "-book1"); } return Uni.createFrom().voidItem(); })); } public boolean processNewAuthor(String authorName) { persistAuthor(authorName); return true; } @Transactional public void persistAuthor(String name) { Author author = new Author(); author.name = name; author.persist(); } private boolean isAlreadyProcessed(Throwable t) { return t instanceof IntegrityConstraintViolationException && t.getCause() != null && t.getCause().getMessage().contains("author_name_key"); } }
I haven't played a lot with this pattern, but I think you need to make up to 3 database transactions using this pattern and I wouldn't use the KafkaTransactions here, but:
@ApplicationScoped
public class NewAuthorConsumer {
@Channel("new-book-out")
Emitter<String> emitter;
@Inject AuthorService authorService;
@Incoming("new-author-in")
Uni<Void> consume(Message<String> message) {
// db tx 1 to create the author
var id = authorService.handleAuthor(message.getPayload());
// db tx 2 to query the author (this is just an example)
var name = authorService.getAuthorName(id);
return emitter.send(name).onFailure()
.recoverWithUni(ex -> {
// failed to send
// db tx 3 to delete the author
authorService.deleteAuthor(id);
message.nack(ex);
})
.onItem().transformToUni(unused -> message.ack());
}
}
Note that I haven't tested this code and also make sure that the commit strategy is manual for this consumer.
@ozangunalp was this related to the work you did on the emitter recently?
Note that database and Kafka transactions have very different semantic. Thus, I would not use them together as it is very likely not going to do what you think (especially when reading the KIP about Kafka transaction (server side defense))
@cescoffier in part, yes. The other part is to be able to run blocking code inside the Kafka transaction.
Note that database and Kafka transactions have very different semantic. Thus, I would not use them together as it is very likely not going to do what you think
I think it is always useful to repeat that. The KIP-890 will bring some stability improvements to Kafka transactions but it is not there yet.
Describe the bug
I'm using a Quarkus service that requires:
All the above steps need to be executed in a single transaction, so for example, if the database fails, the message is nack, or if the emission of the messages fails the database queries are rollback.
In the Quarkus service, we're using the smallrye messaging kafka and the Hibernate ORM extensions. I'm aware that there is an existing section in the Kafka guide to chain transactions between Kafka and Hibernate Reactive: https://quarkus.io/guides/kafka#chaining-kafka-transactions-with-hibernate-reactive-transactions, but nothing about to do that with Hibernate ORM which seems to me like a very important requirement.
I've written a reproducer in https://github.com/Sgitario/quarkus-examples/tree/main/kafka-orm-transactions with several approaches I tried (see NewAuthorConsumer.java).
In the reproducer, the primary key of author is the name, so we can't have two authors with the same name. As part of the initial "import.sql", we have created an author "Jose" without books. Then, we have a resource to send a message to a topic:
And this message will be consumed by the class NewAuthorConsumer that will insert the author using the given name and send another message to another topic to create the books. What I expected here is that when adding another author with the same name "Jose" and a constraint exception is thrown, the messages to the book topic are not sent.
The implementations of the NewAuthorConsumer that I tried:
Using the KafkaTransactions:
As expected, this fails because with:
The error is expected and correct.
This time when adding an existing author, it fails to be inserted but the book message is wrongly sent and hence the book is inserted into the db.
Expected behavior
Spite of the nature of the reactive model of the smallrye messaging extension, it should be a way to chain transactions between kafka and Hibernate ORM. If there is a way to do this, it would need to be documented. If not, I feel this is a very strong limitation, yet I'm not sure if this should be a bug or a feature request.
I also raised the same question in https://github.com/smallrye/smallrye-reactive-messaging/discussions/2609.
Actual behavior
The kafka and the Hiberante ORM transactions are not chained.
How to Reproduce?
1.- git clone https://github.com/Sgitario/quarkus-examples 2.- cd quarkus-examples/kafka-orm-transactions 3.- comment/uncomment/add the approach you would like to use in NewAuthorConsumer 4.- start dev mode:
mvn quarkus:dev
5.- add the existing author:curl -X POST http://localhost:8080/api/authors/Jose
6.- you should see the primary key exception in the service logs:7.- check the books:
curl http://localhost:8080/api/books
And because of the previous exception, the message should have not been sent and there should not be any books. However, the message was wrongly sent and there are books:
Quarkus version or git rev
999-SNAPSHOT
Build tool (ie. output of
mvnw --version
orgradlew --version
)mvn