In the third Article I faced the following text:
Remember the same application code runs on Tomcat, Jetty or Netty.
Currently, the Tomcat and Jetty support is provided on top of Servlet
3.1 asynchronous processing, so it is limited to one request per thread. When the same code runs on the Netty server platform that
constraint is lifted, and the server can dispatch requests
sympathetically to the web client. As long as the client doesn’t
block, everyone is happy. Performance metrics for the netty server and
client probably show similar characteristics, but the Netty server is
not restricted to processing a single request per thread, so it
doesn’t use a large thread pool and we might expect to see some
differences in resource utilization. We will come back to that later
in another article in this series.
First of all I don't see newer article in the series although it was written in 2016. It is clear for me that tomcat has 100 threads by default for handling requests and one thread handle one request in the same time but I don't understand phrase it is limited to one request per thread What does it mean?
Also I would like to know how Netty works for that concrete case(I want to understand difference with Tomcat). Can it handle 2 requests per thread?