1 |
42941
|
claudio.at
|
package eu.dnetlib.msro.workflows.nodes.hadoop;
|
2 |
|
|
|
3 |
|
|
import java.util.List;
|
4 |
|
|
|
5 |
|
|
import eu.dnetlib.msro.workflows.graph.Arc;
|
6 |
|
|
import eu.dnetlib.msro.workflows.procs.Env;
|
7 |
|
|
import eu.dnetlib.rmi.data.hadoop.HadoopService;
|
8 |
|
|
import org.apache.commons.logging.Log;
|
9 |
|
|
import org.apache.commons.logging.LogFactory;
|
10 |
|
|
|
11 |
|
|
public class GetHBaseTableDescriptionJobNode extends AbstractHBaseAdminJobNode {
|
12 |
|
|
|
13 |
|
|
private static final Log log = LogFactory.getLog(GetHBaseTableDescriptionJobNode.class); // NOPMD by marko on 11/24/08 5:02 PM
|
14 |
|
|
|
15 |
|
|
@Override
|
16 |
|
|
protected String execute(final Env env) throws Exception {
|
17 |
|
|
|
18 |
|
|
log.info("getting table description: '" + getTableName() + "' on cluster: '" + getCluster() + "'");
|
19 |
|
|
|
20 |
|
|
final List<String> columns = getServiceLocator().getService(HadoopService.class).describeHbaseTable(getCluster(), getTableName());
|
21 |
|
|
log.debug(String.format("table '%s': " + columns, getTableName()));
|
22 |
|
|
env.setAttribute(getTableColumnsParamName(), asCSV(columns));
|
23 |
|
|
|
24 |
|
|
return Arc.DEFAULT_ARC;
|
25 |
|
|
}
|
26 |
|
|
|
27 |
|
|
}
|