星空网 > 软件开发 > Java

HttpsURLConnection 利用keepAlive特性进行优化一例

最近项目中,遇到一个报错:

java.lang.OutOfMemoryError: unable to create new native thread

报错的场景是:一个消息的群发,群里总共有50多个人,一群发 tomcat 就崩溃了,查看日志报上面的错误。

这个错误的基本情况是,没有内存来创建新的本地线程了,因为java中创建的线程,最终是要对应到本地线程。基本的解决方法,是减小:xss 的数值,页就是减小 java 线程stack 占用内存的大小,一般可以改小到 128k。

关于这个错误网上有一篇博文可以参考:http://sesame.iteye.com/blog/622670

http://stackoverflow.com/ 上也有很多分析好答案。

You have a chance to face the java.lang.OutOfMemoryError: Unable to create new native thread whenever the JVM asks for a new thread from the OS. Whenever the underlying OS cannot allocate a new native thread, this OutOfMemoryError will be thrown. The exact limit for native threads is very platform-dependent thus we recommend to find out those limits by running a test similar to the below example.

重现错误:

	public static void main(String[] args) {		LongAdder adder = new LongAdder();		while(true){		  new Thread(new Runnable(){		    public void run() {		      try {		      	adder.increment();		        Thread.sleep(10000000);		      } catch(InterruptedException e) { 		      			      }    		    }  		  }).start();		  System.out.println(adder.longValue());		}			}

2331
2331
2331
2333
2335
2340

...
2355

Exception in thread "main" java.lang.OutOfMemoryError: unable to create new native thread
at java.lang.Thread.start0(Native Method)
at java.lang.Thread.start(Thread.java:714)
可以看到我电脑大概可以创建2355个线程,在创建,内存就不够了,爆出错误。

 

虽然通过 修改 xss 以及在群发消息时,每一条消息之间增加几秒的间隔,这样把问题解决了。但是问题的本质似乎还是没有解决。

因为 发送消息使用的 http 协议发送的,原代码中 每发送一条消息,就创建一条 HttpUrlConnection,所以这里应该可以进行优化

在官方文档中找到了相关文档:http://docs.oracle.com/javase/8/docs/technotes/guides/net/http-keepalive.html

What is HTTP Persistent Connections?

HTTP persistent connections, also called HTTP keep-alive, or HTTP connection reuse, is the idea of using the same TCP connection to send and receive multiple HTTP requests/responses, as opposed to opening a new one for every single request/response pair. Using persistent connections is very important for improving HTTP performance.

There are several advantages of using persistent connections, including:

  • Network friendly. Less network traffic due to fewer setting up and tearing down of TCP connections.
  • Reduced latency on subsequent request. Due to avoidance of initial TCP handshake
  • Long lasting connections allowing TCP sufficient time to determine the congestion state of the network, thus to react appropriately.

The advantages are even more obvious with HTTPS or HTTP over SSL/TLS. There, persistent connections may reduce the number of costly SSL/TLS handshake to establish security associations, in addition to the initial TCP connection set up.

In HTTP/1.1, persistent connections are the default behavior of any connection. That is, unless otherwise indicated, the client SHOULD assume that the server will maintain a persistent connection, even after error responses from the server. However, the protocol provides means for a client and a server to signal the closing of a TCP connection.

What makes a connection reusable?

Since TCP by its nature is a stream based protocol, in order to reuse an existing connection, the HTTP protocol has to have a way to indicate the end of the previous response and the beginning of the next one. Thus, it is required that all messages on the connection MUST have a self-defined message length (i.e., one not defined by closure of the connection). Self demarcation is achieved by either setting the Content-Length header, or in the case of chunked transfer encoded entity body, each chunk starts with a size, and the response body ends with a special last chunk.

What happens if there are proxy servers in between?

Since persistent connections applies to only one transport link, it is important that proxy servers correctly signal persistent/or-non-persistent connections separately with its clients and the origin servers (or to other proxy servers). From a HTTP client or server's perspective, as far as persistence connection is concerned, the presence or absence of proxy servers is transparent.

What does the current JDK do for Keep-Alive?

The JDK supports both HTTP/1.1 and HTTP/1.0 persistent connections.

When the application finishes reading the response body or when the application calls close() on the InputStream returned by URLConnection.getInputStream(), the JDK's HTTP protocol handler will try to clean up the connection and if successful, put the connection into a connection cache for reuse by future HTTP requests.

The support for HTTP keep-Alive is done transparently. However, it can be controlled by system properties http.keepAlive, and http.maxConnections, as well as by HTTP/1.1 specified request and response headers.

The system properties that control the behavior of Keep-Alive are:

http.keepAlive=<boolean>
default: true

Indicates if keep alive (persistent) connections should be supported.

http.maxConnections=<int>
default: 5

Indicates the maximum number of connections per destination to be kept alive at any given time

HTTP header that influences connection persistence is:

Connection: close

If the "Connection" header is specified with the value "close" in either the request or the response header fields, it indicates that the connection should not be considered 'persistent' after the current request/response is complete.

The current implementation doesn't buffer the response body. Which means that the application has to finish reading the response body or call close() to abandon the rest of the response body, in order for that connection to be reused. Furthermore, current implementation will not try block-reading when cleaning up the connection, meaning if the whole response body is not available, the connection will not be reused.

What's new in JDK 5?

When the application encounters a HTTP 400 or 500 response, it may ignore the IOException and then may issue another HTTP request. In this case, the underlying TCP connection won't be Kept-Alive because the response body is still there to be consumed, so the socket connection is not cleared, therefore not available for reuse. What the application needs to do is callHttpURLConnection.getErrorStream() after catching the IOException , read the response body, then close the stream. However, some existing applications are not doing this. As a result, they do not benefit from persistent connections. To address this problem, we have introduced a workaround.

The workaround involves buffering the response body if the response is >=400, up to a certain amount and within a time limit, thus freeing up the underlying socket connection for reuse. The rationale behind this is that when the server responds with a >=400 error (client error or server error. One example is "404: File Not Found" error), the server usually sends a small response body to explain whom to contact and what to do to recover.

Several new Oracle JDK implementation specific properties are introduced to help clean up the connections after error response from the server.

The major one is:

sun.net.http.errorstream.enableBuffering=<boolean>
default: false

With the above system property set to true (default is false), when the response code is >=400, the HTTP handler will try to buffer the response body. Thus freeing up the underlying socket connection for reuse. Thus, even if the application doesn't call getErrorStream(), read the response body, and then call close(), the underlying socket connection may still be kept-alive and reused.

The following two system properties provide further control to the error stream buffering behavior:

sun.net.http.errorstream.timeout=<int> in millisecond
default: 300 millisecond

sun.net.http.errorstream.bufferSize=<int> in bytes
default: 4096 bytes

What can you do to help with Keep-Alive?

Do not abandon a connection by ignoring the response body. Doing so may results in idle TCP connections. That needs to be garbage collected when they are no longer referenced.

If getInputStream() successfully returns, read the entire response body.

When calling getInputStream() from HttpURLConnection, if an IOException occurs, catch the exception and call getErrorStream() to get the response body (if there is any).

Reading the response body cleans up the connection even if you are not interested in the response content itself. But if the response body is long and you are not interested in the rest of it after seeing the beginning, you can close the InputStream. But you need to be aware that more data could be on its way. Thus the connection may not be cleared for reuse.

Here's a code example that complies to the above recommendation:

try {    URL a = new URL(args[0]);    URLConnection urlc = a.openConnection();    is = conn.getInputStream();    int ret = 0;    while ((ret = is.read(buf)) > 0) {     processBuf(buf);    }    // close the inputstream    is.close();} catch (IOException e) {    try {        respCode = ((HttpURLConnection)conn).getResponseCode();        es = ((HttpURLConnection)conn).getErrorStream();        int ret = 0;        // read the response body        while ((ret = es.read(buf)) > 0) {            processBuf(buf);        }        // close the errorstream        es.close();    } catch(IOException ex) {        // deal with the exception    }}

If you know ahead of time that you won't be interested in the response body, you should issue a HEAD request instead of a GET request. For example when you are only interested in the meta info of the web resource or when testing for its validity, accessibility and recent modification. Here's a code snippet:

URL a = new URL(args[0]);URLConnection urlc = a.openConnection();HttpURLConnection httpc = (HttpURLConnection)urlc;// only interested in the length of the resourcehttpc.setRequestMethod("HEAD");int len = httpc.getContentLength();

Changes in JDK 6

Prior to JDK 6, if an application closes a HTTP InputStream when more than a small amount of data remains to be read, then the connection had to be closed, rather than being cached. Now inJDK 6, the behavior is to read up to 512 Kbytes off the connection in a background thread, thus allowing the connection to be reused. The exact amount of data which may be read is configurable through the http.KeepAlive.remainingData system property.

根据这篇文章的说明,将实行改成使用 keepAlive 的httpUrlConnection来实现,性能得到了很大的提升。
注意,一定要 server 端支持 Keepalive,判断方法:看 http 头部 Connection:keepalive 是否是这个值。
代码:
public static void main(String[] args) {    try {        URL obj = new URL("www.xxx.com");        URLConnection conn = obj.openConnection();        Map<String, List<String>> map = conn.getHeaderFields();           System.out.println("显示响应Header信息...\n");           for (Map.Entry<String, List<String>> entry : map.entrySet()) {            System.out.println("Key : " + entry.getKey() +                      " ,Value : " + entry.getValue());        }//        		Key : Connection ,Value : [keep-alive]           System.out.println("\n使用key获得响应Header信息 \n");        List<String> server = map.get("Server");           if (server == null) {            System.out.println("Key 'Server' is not found!");        } else {            for (String values : server) {                System.out.println(values);            }        }       } catch (Exception e) {        e.printStackTrace();    }	}

大量的HttpUrlConnection、大量的TCP连接,肯定会占用系统内存。改成共用一条HttpUrlConnection/TCP连接,内存占用减少很多。  


 

 

  

 




原标题:HttpsURLConnection 利用keepAlive特性进行优化一例

关键词:URL

URL
*特别声明:以上内容来自于网络收集,著作权属原作者所有,如有侵权,请联系我们: admin#shaoqun.com (#换成@)。

备战黑五网一,3000字教你广告优化重点在哪里!:https://www.ikjzd.com/articles/1592899657134731266
2017年黑五德国站大面积爆单,2018年怎么玩转亚马逊德国站?:https://www.ikjzd.com/articles/1593
收入达历史新高,Coupang三季度营收51亿:https://www.ikjzd.com/articles/1593072431786713090
库存积压、进口下降,美国进口量将出现两年来最低点:https://www.ikjzd.com/articles/1593075816166416386
抢占旺季市场!欧美线下零售商已经动手了……:https://www.ikjzd.com/articles/1593083358825115649
中国空间站助力中国光伏占领欧洲?:https://www.ikjzd.com/articles/1593087226246160385
4月1号,德国《大麻法案》正式生效。新的机会独立站爆品一植物生长灯。:https://www.kjdsnews.com/a/1840745.html
4月1号,德国《大麻法案》正式生效。新的机会独立站爆品一植物生长灯。:https://www.goluckyvip.com/news/214215.html
相关文章
我的浏览记录
最新相关资讯
海外公司注册 | 跨境电商服务平台 | 深圳旅行社 | 东南亚物流