Java tutorial
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.solr.client.solrj.impl; import java.io.IOException; import java.io.InputStream; import java.io.OutputStream; import java.lang.invoke.MethodHandles; import java.nio.charset.StandardCharsets; import java.util.LinkedList; import java.util.Locale; import java.util.Queue; import java.util.Set; import java.util.concurrent.BlockingQueue; import java.util.concurrent.CountDownLatch; import java.util.concurrent.ExecutorService; import java.util.concurrent.LinkedBlockingQueue; import java.util.concurrent.TimeUnit; import org.apache.http.HttpResponse; import org.apache.http.HttpStatus; import org.apache.http.NameValuePair; import org.apache.http.client.HttpClient; import org.apache.http.client.methods.HttpPost; import org.apache.http.entity.ContentProducer; import org.apache.http.entity.EntityTemplate; import org.apache.solr.client.solrj.ResponseParser; import org.apache.solr.client.solrj.SolrClient; import org.apache.solr.client.solrj.SolrRequest; import org.apache.solr.client.solrj.SolrServerException; import org.apache.solr.client.solrj.request.RequestWriter; import org.apache.solr.client.solrj.request.UpdateRequest; import org.apache.solr.client.solrj.util.ClientUtils; import org.apache.solr.common.SolrException; import org.apache.solr.common.SolrException.ErrorCode; import org.apache.solr.common.params.CommonParams; import org.apache.solr.common.params.ModifiableSolrParams; import org.apache.solr.common.params.SolrParams; import org.apache.solr.common.params.UpdateParams; import org.apache.solr.common.util.ExecutorUtil; import org.apache.solr.common.util.IOUtils; import org.apache.solr.common.util.NamedList; import org.apache.solr.common.util.SolrjNamedThreadFactory; import org.apache.solr.common.util.Utils; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.slf4j.MDC; /** * ConcurrentUpdateSolrClient buffers all added documents and writes * them into open HTTP connections. This class is thread safe. * * Params from {@link UpdateRequest} are converted to http request * parameters. When params change between UpdateRequests a new HTTP * request is started. * * Although any SolrClient request can be made with this implementation, it is * only recommended to use ConcurrentUpdateSolrClient with /update * requests. The class {@link HttpSolrClient} is better suited for the * query interface. */ public class ConcurrentUpdateSolrClient extends SolrClient { private static final long serialVersionUID = 1L; private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass()); private HttpSolrClient client; final BlockingQueue<Update> queue; final ExecutorService scheduler; final Queue<Runner> runners; volatile CountDownLatch lock = null; // used to block everything final int threadCount; boolean shutdownExecutor = false; int pollQueueTime = 250; private final boolean streamDeletes; private boolean internalHttpClient; private volatile Integer connectionTimeout; private volatile Integer soTimeout; /** * Uses an internally managed HttpClient instance. * * @param solrServerUrl * The Solr server URL * @param queueSize * The buffer size before the documents are sent to the server * @param threadCount * The number of background threads used to empty the queue * * @deprecated use {@link Builder} instead. */ @Deprecated public ConcurrentUpdateSolrClient(String solrServerUrl, int queueSize, int threadCount) { this(solrServerUrl, null, queueSize, threadCount); shutdownExecutor = true; internalHttpClient = true; } /** * @deprecated use {@link Builder} instead. */ @Deprecated public ConcurrentUpdateSolrClient(String solrServerUrl, HttpClient client, int queueSize, int threadCount) { this(solrServerUrl, client, queueSize, threadCount, ExecutorUtil.newMDCAwareCachedThreadPool(new SolrjNamedThreadFactory("concurrentUpdateScheduler"))); shutdownExecutor = true; } /** * Uses the supplied HttpClient to send documents to the Solr server. * * @deprecated use {@link Builder} instead. */ @Deprecated public ConcurrentUpdateSolrClient(String solrServerUrl, HttpClient client, int queueSize, int threadCount, ExecutorService es) { this(solrServerUrl, client, queueSize, threadCount, es, false); } /** * Uses the supplied HttpClient to send documents to the Solr server. * * @deprecated use {@link Builder} instead. This will soon be a * protected method, and will only be available for use in implementing subclasses. */ @Deprecated public ConcurrentUpdateSolrClient(String solrServerUrl, HttpClient client, int queueSize, int threadCount, ExecutorService es, boolean streamDeletes) { this.internalHttpClient = (client == null); this.client = new HttpSolrClient.Builder(solrServerUrl).withHttpClient(client).build(); this.client.setFollowRedirects(false); queue = new LinkedBlockingQueue<>(queueSize); this.threadCount = threadCount; runners = new LinkedList<>(); this.streamDeletes = streamDeletes; if (es != null) { scheduler = es; shutdownExecutor = false; } else { scheduler = ExecutorUtil .newMDCAwareCachedThreadPool(new SolrjNamedThreadFactory("concurrentUpdateScheduler")); shutdownExecutor = true; } } public Set<String> getQueryParams() { return this.client.getQueryParams(); } /** * Expert Method. * @param queryParams set of param keys to only send via the query string */ public void setQueryParams(Set<String> queryParams) { this.client.setQueryParams(queryParams); } /** * Opens a connection and sends everything... */ class Runner implements Runnable { @Override public void run() { log.debug("starting runner: {}", this); // This loop is so we can continue if an element was added to the queue after the last runner exited. for (;;) { try { sendUpdateStream(); } catch (Throwable e) { if (e instanceof OutOfMemoryError) { throw (OutOfMemoryError) e; } handleError(e); } finally { synchronized (runners) { // check to see if anything else was added to the queue if (runners.size() == 1 && !queue.isEmpty() && !scheduler.isShutdown()) { // If there is something else to process, keep last runner alive by staying in the loop. } else { runners.remove(this); if (runners.isEmpty()) { // notify anyone waiting in blockUntilFinished runners.notifyAll(); } break; } } } } log.debug("finished: {}", this); } // // Pull from the queue multiple times and streams over a single connection. // Exits on exception, interruption, or an empty queue to pull from. // void sendUpdateStream() throws Exception { while (!queue.isEmpty()) { HttpPost method = null; HttpResponse response = null; InputStream rspBody = null; try { final Update update = queue.poll(pollQueueTime, TimeUnit.MILLISECONDS); if (update == null) break; String contentType = client.requestWriter.getUpdateContentType(); final boolean isXml = ClientUtils.TEXT_XML.equals(contentType); final ModifiableSolrParams origParams = new ModifiableSolrParams( update.getRequest().getParams()); EntityTemplate template = new EntityTemplate(new ContentProducer() { @Override public void writeTo(OutputStream out) throws IOException { try { if (isXml) { out.write("<stream>".getBytes(StandardCharsets.UTF_8)); // can be anything } Update upd = update; while (upd != null) { UpdateRequest req = upd.getRequest(); SolrParams currentParams = new ModifiableSolrParams(req.getParams()); if (!origParams.toNamedList().equals(currentParams.toNamedList())) { queue.add(upd); // params are different, push back to queue break; } client.requestWriter.write(req, out); if (isXml) { // check for commit or optimize SolrParams params = req.getParams(); if (params != null) { String fmt = null; if (params.getBool(UpdateParams.OPTIMIZE, false)) { fmt = "<optimize waitSearcher=\"%s\" />"; } else if (params.getBool(UpdateParams.COMMIT, false)) { fmt = "<commit waitSearcher=\"%s\" />"; } if (fmt != null) { byte[] content = String.format(Locale.ROOT, fmt, params.getBool(UpdateParams.WAIT_SEARCHER, false) + "") .getBytes(StandardCharsets.UTF_8); out.write(content); } } } out.flush(); if (pollQueueTime > 0 && threadCount == 1 && req.isLastDocInBatch()) { // no need to wait to see another doc in the queue if we've hit the last doc in a batch upd = queue.poll(0, TimeUnit.MILLISECONDS); } else { upd = queue.poll(pollQueueTime, TimeUnit.MILLISECONDS); } } if (isXml) { out.write("</stream>".getBytes(StandardCharsets.UTF_8)); } } catch (InterruptedException e) { Thread.currentThread().interrupt(); log.warn("", e); } } }); // The parser 'wt=' and 'version=' params are used instead of the // original params ModifiableSolrParams requestParams = new ModifiableSolrParams(origParams); requestParams.set(CommonParams.WT, client.parser.getWriterType()); requestParams.set(CommonParams.VERSION, client.parser.getVersion()); String basePath = client.getBaseURL(); if (update.getCollection() != null) basePath += "/" + update.getCollection(); method = new HttpPost(basePath + "/update" + requestParams.toQueryString()); org.apache.http.client.config.RequestConfig.Builder requestConfigBuilder = HttpClientUtil .createDefaultRequestConfigBuilder(); if (soTimeout != null) { requestConfigBuilder.setSocketTimeout(soTimeout); } if (connectionTimeout != null) { requestConfigBuilder.setConnectTimeout(connectionTimeout); } method.setConfig(requestConfigBuilder.build()); method.setEntity(template); method.addHeader("User-Agent", HttpSolrClient.AGENT); method.addHeader("Content-Type", contentType); response = client.getHttpClient().execute(method, HttpClientUtil.createNewHttpClientRequestContext()); rspBody = response.getEntity().getContent(); int statusCode = response.getStatusLine().getStatusCode(); if (statusCode != HttpStatus.SC_OK) { StringBuilder msg = new StringBuilder(); msg.append(response.getStatusLine().getReasonPhrase()); msg.append("\n\n\n\n"); msg.append("request: ").append(method.getURI()); SolrException solrExc = new SolrException(ErrorCode.getErrorCode(statusCode), msg.toString()); // parse out the metadata from the SolrException try { String encoding = "UTF-8"; // default if (response.getEntity().getContentType().getElements().length > 0) { NameValuePair param = response.getEntity().getContentType().getElements()[0] .getParameterByName("charset"); if (param != null) { encoding = param.getValue(); } } NamedList<Object> resp = client.parser.processResponse(rspBody, encoding); NamedList<Object> error = (NamedList<Object>) resp.get("error"); if (error != null) { solrExc.setMetadata((NamedList<String>) error.get("metadata")); } } catch (Exception exc) { // don't want to fail to report error if parsing the response fails log.warn("Failed to parse error response from " + client.getBaseURL() + " due to: " + exc); } handleError(solrExc); } else { onSuccess(response); } } finally { try { if (response != null) { Utils.consumeFully(response.getEntity()); } } catch (Exception e) { log.error("Error consuming and closing http response stream.", e); } } } } } // *must* be called with runners monitor held, e.g. synchronized(runners){ addRunner() } private void addRunner() { MDC.put("ConcurrentUpdateSolrClient.url", client.getBaseURL()); try { Runner r = new Runner(); runners.add(r); scheduler.execute(r); // this can throw an exception if the scheduler has been shutdown, but that should be fine. } finally { MDC.remove("ConcurrentUpdateSolrClient.url"); } } /** * Class representing an UpdateRequest and an optional collection. */ class Update { UpdateRequest request; String collection; /** * * @param request the update request. * @param collection The collection, can be null. */ public Update(UpdateRequest request, String collection) { this.request = request; this.collection = collection; } /** * @return the update request. */ public UpdateRequest getRequest() { return request; } public void setRequest(UpdateRequest request) { this.request = request; } /** * @return the collection, can be null. */ public String getCollection() { return collection; } public void setCollection(String collection) { this.collection = collection; } } @Override public NamedList<Object> request(final SolrRequest request, String collection) throws SolrServerException, IOException { if (!(request instanceof UpdateRequest)) { return client.request(request, collection); } UpdateRequest req = (UpdateRequest) request; // this happens for commit... if (streamDeletes) { if ((req.getDocuments() == null || req.getDocuments().isEmpty()) && (req.getDeleteById() == null || req.getDeleteById().isEmpty()) && (req.getDeleteByIdMap() == null || req.getDeleteByIdMap().isEmpty())) { if (req.getDeleteQuery() == null) { blockUntilFinished(); return client.request(request, collection); } } } else { if ((req.getDocuments() == null || req.getDocuments().isEmpty())) { blockUntilFinished(); return client.request(request, collection); } } SolrParams params = req.getParams(); if (params != null) { // check if it is waiting for the searcher if (params.getBool(UpdateParams.WAIT_SEARCHER, false)) { log.info("blocking for commit/optimize"); blockUntilFinished(); // empty the queue return client.request(request, collection); } } try { CountDownLatch tmpLock = lock; if (tmpLock != null) { tmpLock.await(); } Update update = new Update(req, collection); boolean success = queue.offer(update); for (;;) { synchronized (runners) { // see if queue is half full and we can add more runners // special case: if only using a threadCount of 1 and the queue // is filling up, allow 1 add'l runner to help process the queue if (runners.isEmpty() || (queue.remainingCapacity() < queue.size() && runners.size() < threadCount)) { // We need more runners, so start a new one. addRunner(); } else { // break out of the retry loop if we added the element to the queue // successfully, *and* // while we are still holding the runners lock to prevent race // conditions. if (success) break; } } // Retry to add to the queue w/o the runners lock held (else we risk // temporary deadlock) // This retry could also fail because // 1) existing runners were not able to take off any new elements in the // queue // 2) the queue was filled back up since our last try // If we succeed, the queue may have been completely emptied, and all // runners stopped. // In all cases, we should loop back to the top to see if we need to // start more runners. // if (!success) { success = queue.offer(update, 100, TimeUnit.MILLISECONDS); } } } catch (InterruptedException e) { log.error("interrupted", e); throw new IOException(e.getLocalizedMessage()); } // RETURN A DUMMY result NamedList<Object> dummy = new NamedList<>(); dummy.add("NOTE", "the request is processed in a background stream"); return dummy; } public synchronized void blockUntilFinished() { lock = new CountDownLatch(1); try { synchronized (runners) { // NOTE: if the executor is shut down, runners may never become empty (a scheduled task may never be run, // which means it would never remove itself from the runners list. This is why we don't wait forever // and periodically check if the scheduler is shutting down. while (!runners.isEmpty()) { try { runners.wait(250); } catch (InterruptedException e) { Thread.interrupted(); } if (scheduler.isShutdown()) break; // Need to check if the queue is empty before really considering this is finished (SOLR-4260) int queueSize = queue.size(); if (queueSize > 0 && runners.isEmpty()) { // TODO: can this still happen? log.warn("No more runners, but queue still has " + queueSize + " adding more runners to process remaining requests on queue"); addRunner(); } } } } finally { lock.countDown(); lock = null; } } public void handleError(Throwable ex) { log.error("error", ex); } /** * Intended to be used as an extension point for doing post processing after a request completes. */ public void onSuccess(HttpResponse resp) { // no-op by design, override to add functionality } @Override public void close() { if (internalHttpClient) IOUtils.closeQuietly(client); if (shutdownExecutor) { scheduler.shutdown(); try { if (!scheduler.awaitTermination(60, TimeUnit.SECONDS)) { scheduler.shutdownNow(); if (!scheduler.awaitTermination(60, TimeUnit.SECONDS)) log.error("ExecutorService did not terminate"); } } catch (InterruptedException ie) { scheduler.shutdownNow(); Thread.currentThread().interrupt(); } } } public void setConnectionTimeout(int timeout) { this.connectionTimeout = timeout; } /** * set soTimeout (read timeout) on the underlying HttpConnectionManager. This is desirable for queries, but probably * not for indexing. */ public void setSoTimeout(int timeout) { this.soTimeout = timeout; } public void shutdownNow() { if (internalHttpClient) IOUtils.closeQuietly(client); if (shutdownExecutor) { scheduler.shutdownNow(); // Cancel currently executing tasks try { if (!scheduler.awaitTermination(30, TimeUnit.SECONDS)) log.error("ExecutorService did not terminate"); } catch (InterruptedException ie) { scheduler.shutdownNow(); Thread.currentThread().interrupt(); } } } public void setParser(ResponseParser responseParser) { client.setParser(responseParser); } /** * @param pollQueueTime time for an open connection to wait for updates when * the queue is empty. */ public void setPollQueueTime(int pollQueueTime) { this.pollQueueTime = pollQueueTime; } public void setRequestWriter(RequestWriter requestWriter) { client.setRequestWriter(requestWriter); } /** * Constructs {@link ConcurrentUpdateSolrClient} instances from provided configuration. */ public static class Builder { private String baseSolrUrl; private HttpClient httpClient; private int queueSize; private int threadCount; private ExecutorService executorService; private boolean streamDeletes; /** * Create a Builder object, based on the provided Solr URL. * * @param baseSolrUrl the base URL of the Solr server that will be targeted by any created clients. */ public Builder(String baseSolrUrl) { this.baseSolrUrl = baseSolrUrl; } /** * Provides a {@link HttpClient} for the builder to use when creating clients. */ public Builder withHttpClient(HttpClient httpClient) { this.httpClient = httpClient; return this; } /** * The number of documents to batch together before sending to Solr. */ public Builder withQueueSize(int queueSize) { if (queueSize <= 0) { throw new IllegalArgumentException("queueSize must be a positive integer."); } this.queueSize = queueSize; return this; } /** * The number of threads used to empty {@link ConcurrentUpdateSolrClient}s queue. */ public Builder withThreadCount(int threadCount) { if (threadCount <= 0) { throw new IllegalArgumentException("threadCount must be a positive integer."); } this.threadCount = threadCount; return this; } /** * Provides the {@link ExecutorService} for clients to use when servicing requests. */ public Builder withExecutorService(ExecutorService executorService) { this.executorService = executorService; return this; } /** * Configures created clients to always stream delete requests. */ public Builder alwaysStreamDeletes() { this.streamDeletes = true; return this; } /** * Configures created clients to not stream delete requests. */ public Builder neverStreamDeletes() { this.streamDeletes = false; return this; } /** * Create a {@link ConcurrentUpdateSolrClient} based on the provided configuration options. */ public ConcurrentUpdateSolrClient build() { if (baseSolrUrl == null) { throw new IllegalArgumentException("Cannot create HttpSolrClient without a valid baseSolrUrl!"); } return new ConcurrentUpdateSolrClient(baseSolrUrl, httpClient, queueSize, threadCount, executorService, streamDeletes); } } }