public class DagEngine extends BaseEngine
USE_XCOMMAND, user| Constructor and Description |
|---|
DagEngine()
Create a system Dag engine, with no user and no group.
|
DagEngine(String user)
Create a Dag engine to perform operations on behave of a user.
|
| Modifier and Type | Method and Description |
|---|---|
void |
change(String jobId,
String changeValue)
Change a coordinator job.
|
void |
changeSLA(String id,
String actions,
String dates,
String childIds,
String newParams)
Change SLA properties for job
|
void |
disableSLAAlert(String id,
String actions,
String dates,
String childIds)
Disable SLA alert for job
|
String |
dryRunSubmit(org.apache.hadoop.conf.Configuration conf)
Dry run a job; like
BaseEngine.submitJob(org.apache.hadoop.conf.Configuration, boolean) but doesn't actually execute
the job. |
void |
enableSLAAlert(String id,
String actions,
String dates,
String childIds)
Enable SLA alert for job
|
CoordinatorJob |
getCoordJob(String jobId)
Return the info about a coord job.
|
CoordinatorJob |
getCoordJob(String jobId,
String filter,
int start,
int length,
boolean desc)
Return the info about a coord job with actions subset.
|
String |
getDefinition(String jobId)
Return the a job definition.
|
WorkflowJob |
getJob(String jobId)
Return the info about a job.
|
WorkflowJob |
getJob(String jobId,
int start,
int length)
Return the info about a job with actions subset.
|
String |
getJobIdForExternalId(String externalId)
Return the workflow Job ID for an external ID.
|
WorkflowsInfo |
getJobs(String filter,
int start,
int len)
Return the info about a set of jobs.
|
String |
getJobStatus(String jobId)
Return the status for a Job ID
|
WorkflowActionBean |
getWorkflowAction(String actionId) |
List<Map<String,String>> |
getWorkflowActionRetries(String actionId)
Gets the workflow action retries.
|
void |
kill(String jobId)
Kill a job.
|
WorkflowsInfo |
killJobs(String filter,
int start,
int len)
return the jobs that've been killed
|
protected Map<String,List<String>> |
parseFilter(String filter)
Validate a jobs filter.
|
void |
processCallback(String actionId,
String externalStatus,
Properties actionData)
Process an action callback.
|
void |
reRun(String jobId,
org.apache.hadoop.conf.Configuration conf)
Rerun a job.
|
void |
resume(String jobId)
Resume a job.
|
WorkflowsInfo |
resumeJobs(String filter,
int start,
int len)
return the jobs that've been resumed
|
void |
start(String jobId)
Start a job.
|
protected void |
streamJobLog(XLogStreamer logStreamer,
String jobId,
Writer writer)
Stream job log.
|
String |
submitHttpJob(org.apache.hadoop.conf.Configuration conf,
String jobType)
Submit a pig/hive/mapreduce job through HTTP.
|
String |
submitJob(org.apache.hadoop.conf.Configuration conf,
boolean startJob)
Submit a workflow job.
|
String |
submitJobFromCoordinator(org.apache.hadoop.conf.Configuration conf,
String parentId)
Submit a workflow through a coordinator.
|
void |
suspend(String jobId)
Suspend a job.
|
WorkflowsInfo |
suspendJobs(String filter,
int start,
int len)
return the jobs that've been suspended
|
getJMSTopicName, getUser, streamAuditLog, streamErrorLog, streamLogpublic DagEngine()
public String submitJob(org.apache.hadoop.conf.Configuration conf, boolean startJob) throws DagEngineException
It validates configuration properties.
submitJob in class BaseEngineconf - job configuration.startJob - indicates if the job should be started or not.DagEngineException - thrown if the job could not be created.public String submitJobFromCoordinator(org.apache.hadoop.conf.Configuration conf, String parentId) throws DagEngineException
conf - job confparentId - parent of workflowDagEngineException - if the job can't be submittedpublic String submitHttpJob(org.apache.hadoop.conf.Configuration conf, String jobType) throws DagEngineException
It validates configuration properties.
conf - job configuration.jobType - job type - can be "pig", "hive", "sqoop" or "mapreduce".DagEngineException - thrown if the job could not be created.public void start(String jobId) throws DagEngineException
start in class BaseEnginejobId - job Id.DagEngineException - thrown if the job could not be started.public void resume(String jobId) throws DagEngineException
resume in class BaseEnginejobId - job Id.DagEngineException - thrown if the job could not be resumed.public void suspend(String jobId) throws DagEngineException
suspend in class BaseEnginejobId - job Id.DagEngineException - thrown if the job could not be suspended.public void kill(String jobId) throws DagEngineException
kill in class BaseEnginejobId - job Id.DagEngineException - thrown if the job could not be killed.public void change(String jobId, String changeValue) throws DagEngineException
BaseEnginechange in class BaseEnginejobId - job Id.changeValue - change value.DagEngineExceptionpublic void reRun(String jobId, org.apache.hadoop.conf.Configuration conf) throws DagEngineException
reRun in class BaseEnginejobId - job Id to rerun.conf - configuration information for the rerun.DagEngineException - thrown if the job could not be rerun.public void processCallback(String actionId, String externalStatus, Properties actionData) throws DagEngineException
actionId - the action Id.externalStatus - the action external status.actionData - the action output data, null if none.DagEngineException - thrown if the callback could not be processed.public WorkflowJob getJob(String jobId) throws DagEngineException
getJob in class BaseEnginejobId - job Id.DagEngineException - thrown if the job info could not be obtained.public WorkflowJob getJob(String jobId, int start, int length) throws DagEngineException
getJob in class BaseEnginejobId - job Idstart - starting from this index in the list of actions belonging to the joblength - number of actions to be returnedDagEngineException - thrown if the job info could not be obtained.public String getDefinition(String jobId) throws DagEngineException
getDefinition in class BaseEnginejobId - job Id.DagEngineException - thrown if the job definition could no be obtained.protected void streamJobLog(XLogStreamer logStreamer, String jobId, Writer writer) throws IOException, DagEngineException
BaseEnginestreamJobLog in class BaseEnginelogStreamer - the log streamerjobId - the job idwriter - the writerIOException - Signals that an I/O exception has occurred.DagEngineExceptionprotected Map<String,List<String>> parseFilter(String filter) throws DagEngineException
filter - filter to validate.DagEngineException - thrown if the filter is invalid.public WorkflowsInfo getJobs(String filter, int start, int len) throws DagEngineException
filter - job filter. Refer to the OozieClient for the filter syntax.start - offset, base 1.len - number of jobs to return.DagEngineException - thrown if the jobs info could not be obtained.public String getJobIdForExternalId(String externalId) throws DagEngineException
This is reverse lookup for recovery purposes.
getJobIdForExternalId in class BaseEngineexternalId - external ID provided at job submission time.null if none.DagEngineException - thrown if the lookup could not be done.public CoordinatorJob getCoordJob(String jobId) throws BaseEngineException
BaseEnginegetCoordJob in class BaseEnginejobId - job Id.BaseEngineException - thrown if the job info could not be obtained.public CoordinatorJob getCoordJob(String jobId, String filter, int start, int length, boolean desc) throws BaseEngineException
BaseEnginegetCoordJob in class BaseEnginejobId - job Id.filter - the status filterstart - starting from this index in the list of actions belonging to the joblength - number of actions to be returneddesc - true if actions are sorted in a descending order of nominal time, false if asc orderBaseEngineException - thrown if the job info could not be obtained.public WorkflowActionBean getWorkflowAction(String actionId) throws BaseEngineException
BaseEngineExceptionpublic List<Map<String,String>> getWorkflowActionRetries(String actionId) throws BaseEngineException
actionId - the action idBaseEngineException - the base engine exceptionpublic String dryRunSubmit(org.apache.hadoop.conf.Configuration conf) throws BaseEngineException
BaseEngineBaseEngine.submitJob(org.apache.hadoop.conf.Configuration, boolean) but doesn't actually execute
the job.
It validates configuration properties.
dryRunSubmit in class BaseEngineconf - job configuration.BaseEngineException - thrown if there was a problem doing the dryrunpublic String getJobStatus(String jobId) throws DagEngineException
getJobStatus in class BaseEnginejobId - job Id.DagEngineException - thrown if the job's status could not be obtainedpublic void enableSLAAlert(String id, String actions, String dates, String childIds) throws BaseEngineException
BaseEngineenableSLAAlert in class BaseEngineid - job IDactions - list of actionsdates - dateschildIds - child IDsBaseEngineException - thrown if SLA alert could not be enabledpublic void disableSLAAlert(String id, String actions, String dates, String childIds) throws BaseEngineException
BaseEnginedisableSLAAlert in class BaseEngineid - job IDactions - list of actionsdates - dateschildIds - child IDsBaseEngineException - thrown if SLA alert could not be disabledpublic void changeSLA(String id, String actions, String dates, String childIds, String newParams) throws BaseEngineException
BaseEnginechangeSLA in class BaseEngineid - job IDactions - list of actionsdates - dateschildIds - child IDsnewParams - parameters toBaseEngineException - thrown if SLA alert could not be enabledpublic WorkflowsInfo killJobs(String filter, int start, int len) throws DagEngineException
filter - Jobs that satisfy the filter will be killedstart - start index in the database of jobslen - maximum number of jobs that will be killedDagEngineException - if the jobs could not be killedpublic WorkflowsInfo suspendJobs(String filter, int start, int len) throws DagEngineException
filter - Filter for jobs that will be suspended, can be name, user, group, status, id or combination of anystart - Offset for the jobs that will be suspendedlen - maximum number of jobs that will be suspendedDagEngineException - if the jobs cloud not be suspendedpublic WorkflowsInfo resumeJobs(String filter, int start, int len) throws DagEngineException
filter - Filter for jobs that will be resumed, can be name, user, group, status, id or combination of anystart - Offset for the jobs that will be resumedlen - maximum number of jobs that will be resumedDagEngineException - if the jobs cloud not be resumedCopyright © 2018 Apache Software Foundation. All rights reserved.