net.peacesoft.nutch.crawl.ReSolrWriter.java Source code

Java tutorial

Introduction

Here is the source code for net.peacesoft.nutch.crawl.ReSolrWriter.java

Source

/*
 * Licensed to the Apache Software Foundation (ASF) under one or more
 * contributor license agreements.  See the NOTICE file distributed with
 * this work for additional information regarding copyright ownership.
 * The ASF licenses this file to You under the Apache License, Version 2.0
 * (the "License"); you may not use this file except in compliance with
 * the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package net.peacesoft.nutch.crawl;

import java.io.File;
import java.io.FileNotFoundException;
import java.io.FileOutputStream;
import java.io.IOException;
import java.util.ArrayList;
import java.util.Date;
import java.util.List;
import java.util.Map.Entry;
import org.apache.commons.codec.binary.Base64;
import org.apache.commons.httpclient.HttpClient;
import org.apache.commons.httpclient.HttpVersion;
import org.apache.commons.httpclient.methods.PostMethod;
import org.apache.hadoop.mapred.JobConf;
import org.apache.nutch.indexer.IndexerMapReduce;
import org.apache.nutch.indexer.NutchDocument;
import org.apache.nutch.indexer.NutchField;
import org.apache.nutch.indexer.NutchIndexWriter;
import org.apache.nutch.indexer.solr.SolrConstants;
import org.apache.nutch.indexer.solr.SolrMappingReader;
import org.apache.nutch.indexer.solr.SolrUtils;
import org.apache.nutch.indexer.solr.SolrWriter;
import org.apache.solr.client.solrj.SolrServer;
import org.apache.solr.client.solrj.SolrServerException;
import org.apache.solr.client.solrj.request.UpdateRequest;
import org.apache.solr.common.SolrInputDocument;
import org.apache.solr.common.params.ModifiableSolrParams;
import org.apache.solr.common.util.DateUtil;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

public class ReSolrWriter implements NutchIndexWriter {

    public static final Logger LOG = LoggerFactory.getLogger(SolrWriter.class);
    private SolrServer solr;
    private SolrMappingReader solrMapping;
    private ModifiableSolrParams params;
    private final List<SolrInputDocument> inputDocs = new ArrayList<SolrInputDocument>();
    private int commitSize;
    private int numDeletes = 0;
    private boolean delete = false;

    public void open(JobConf job, String name) throws IOException {
        SolrServer server = SolrUtils.getCommonsHttpSolrServer(job);
        init(server, job);
    }

    // package protected for tests
    void init(SolrServer server, JobConf job) throws IOException {
        solr = server;
        commitSize = job.getInt(SolrConstants.COMMIT_SIZE, 1000);
        solrMapping = SolrMappingReader.getInstance(job);
        delete = job.getBoolean(IndexerMapReduce.INDEXER_DELETE, false);
        // parse optional params
        params = new ModifiableSolrParams();
        String paramString = job.get(SolrConstants.PARAMS);
        if (paramString != null) {
            String[] values = paramString.split("&");
            for (String v : values) {
                String[] kv = v.split("=");
                if (kv.length < 2) {
                    continue;
                }
                params.add(kv[0], kv[1]);
            }
        }
    }

    public void delete(String key) throws IOException {
        if (delete) {
            try {
                solr.deleteById(key);
                numDeletes++;
            } catch (final SolrServerException e) {
                throw makeIOException(e);
            }
        }
    }

    public void write(NutchDocument doc) throws IOException {
        final SolrInputDocument inputDoc = new SolrInputDocument();
        for (final Entry<String, NutchField> e : doc) {
            for (final Object val : e.getValue().getValues()) {
                // normalise the string representation for a Date
                Object val2 = val;

                if (val instanceof Date) {
                    val2 = DateUtil.getThreadLocalDateFormat().format(val);
                }

                if (e.getKey().equals("content")) {
                    val2 = SolrUtils.stripNonCharCodepoints((String) val);
                }

                inputDoc.addField(solrMapping.mapKey(e.getKey()), val2, e.getValue().getWeight());
                String sCopy = solrMapping.mapCopyKey(e.getKey());
                if (sCopy != e.getKey()) {
                    inputDoc.addField(sCopy, val);
                }
            }
        }

        postHttp("http://beta2.chodientu.vn/crawler/new", doc);

        inputDoc.setDocumentBoost(doc.getWeight());
        inputDocs.add(inputDoc);
        if (inputDocs.size() + numDeletes >= commitSize) {
            try {
                LOG.info("Indexing " + Integer.toString(inputDocs.size()) + " documents");
                LOG.info("Deleting " + Integer.toString(numDeletes) + " documents");
                numDeletes = 0;
                UpdateRequest req = new UpdateRequest();
                req.add(inputDocs);
                req.setParams(params);
                req.process(solr);
            } catch (final SolrServerException e) {
                throw makeIOException(e);
            }
            inputDocs.clear();
        }
    }

    public void close() throws IOException {
        try {
            if (!inputDocs.isEmpty()) {
                LOG.info("Indexing " + Integer.toString(inputDocs.size()) + " documents");
                if (numDeletes > 0) {
                    LOG.info("Deleting " + Integer.toString(numDeletes) + " documents");
                }
                UpdateRequest req = new UpdateRequest();
                req.add(inputDocs);
                req.setParams(params);
                req.process(solr);
                inputDocs.clear();
            }
        } catch (final SolrServerException e) {
            throw makeIOException(e);
        }
    }

    public void postHttp(String httpServer, NutchDocument doc) {
        try {
            HttpClient client = new HttpClient();
            PostMethod method = new PostMethod(httpServer);
            Object tmp = doc.getFieldValue("myid");
            if (LOG.isInfoEnabled()) {
                LOG.info("Post content id " + tmp.toString() + "to http server: " + httpServer);
            }
            if (tmp != null) {
                method.addParameter("CrawlerContent[id]", tmp.toString());
            }
            tmp = doc.getFieldValue("title");
            if (tmp != null) {
                method.addParameter("CrawlerContent[title]", Base64.encodeBase64String(tmp.toString().getBytes()));
            }
            tmp = doc.getFieldValue("categoryId");
            if (tmp != null) {
                method.addParameter("CrawlerContent[categoryId]",
                        Base64.encodeBase64String(tmp.toString().getBytes()));
            }
            tmp = doc.getFieldValue("categoryChildId");
            if (tmp != null) {
                method.addParameter("CrawlerContent[categoryChildId]",
                        Base64.encodeBase64String(tmp.toString().getBytes()));
            }
            tmp = doc.getFieldValue("content");
            if (tmp != null) {
                method.addParameter("CrawlerContent[content]",
                        Base64.encodeBase64String(tmp.toString().getBytes()));
            }
            tmp = doc.getFieldValue("contentHtml");
            if (tmp != null) {
                method.addParameter("CrawlerContent[contentHtml]", tmp.toString());
            }
            tmp = doc.getFieldValue("location");
            if (tmp != null) {
                method.addParameter("CrawlerContent[Location]",
                        Base64.encodeBase64String(tmp.toString().getBytes()));
            }
            tmp = doc.getFieldValue("tstamp");
            if (tmp != null) {
                method.addParameter("CrawlerContent[createDate]",
                        Base64.encodeBase64String(tmp.toString().getBytes()));
            }
            tmp = doc.getFieldValue("domain");
            if (tmp != null) {
                method.addParameter("CrawlerContent[domain]", tmp.toString());
            }
            tmp = doc.getFieldValue("url");
            if (tmp != null) {
                method.addParameter("CrawlerContent[url]", Base64.encodeBase64String(tmp.toString().getBytes()));
            }
            tmp = doc.getFieldValue("mobile");
            if (tmp != null) {
                method.addParameter("CrawlerContent[mobile]", tmp.toString());
            }
            tmp = doc.getFieldValue("address");
            if (tmp != null) {
                method.addParameter("CrawlerContent[address]",
                        Base64.encodeBase64String(tmp.toString().getBytes()));
            }
            client.executeMethod(method);
            byte[] data = method.getResponseBody();
            toFile(doc.getFieldValue("segment").toString() + ".html", data);
        } catch (Exception ex) {
            LOG.warn("Error when post data to server: " + httpServer, ex);
        }
    }

    public static void toFile(String fileName, byte[] data) {
        FileOutputStream fos = null;
        try {
            File f = new File(fileName);
            fos = new FileOutputStream(f);
            fos.write(data);
            fos.flush();
            fos.close();
        } catch (FileNotFoundException ex) {
        } catch (Exception ex) {
            System.out.println("Loi khi xuat ra file: " + fileName);
        } finally {
            try {
                fos.close();
            } catch (Exception ex) {
            }
        }
    }

    public static IOException makeIOException(SolrServerException e) {
        final IOException ioe = new IOException();
        ioe.initCause(e);
        return ioe;
    }
}