Created
September 22, 2015 11:54
-
-
Save natea/4dde0f50112ec2508a30 to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
2015-09-22 11:51:34,841 INFO 12850 [luigi-interface] hadoop.py:234 - FAILED: SemanticException [Error 10001]: Line 25:16 Table not found 'auth_userprofile' | |
2015-09-22 11:51:34,980 ERROR 12850 [luigi-interface] worker.py:304 - [pid 12850] Worker Worker(salt=039843096, host=precise64, username=hadoop, pid=12850) failed HiveTableFromParameterQueryTask(overwrite=False, warehouse_path=hdfs://localhost:9000/edx-analytics-pipeline/warehouse/, insert_query= | |
SELECT | |
ce.date, | |
ce.course_id, | |
CASE p.level_of_education | |
WHEN 'el' THEN 'primary' | |
WHEN 'jhs' THEN 'junior_secondary' | |
WHEN 'hs' THEN 'secondary' | |
WHEN 'a' THEN 'associates' | |
WHEN 'b' THEN 'bachelors' | |
WHEN 'm' THEN 'masters' | |
WHEN 'p' THEN 'doctorate' | |
WHEN 'p_se' THEN 'doctorate' | |
WHEN 'p_oth' THEN 'doctorate' | |
WHEN 'none' THEN 'none' | |
WHEN 'other' THEN 'other' | |
ELSE NULL | |
END, | |
SUM(ce.at_end), | |
COUNT(ce.user_id) | |
FROM course_enrollment ce | |
LEFT OUTER JOIN auth_userprofile p ON p.user_id = ce.user_id | |
GROUP BY | |
ce.date, | |
ce.course_id, | |
CASE p.level_of_education | |
WHEN 'el' THEN 'primary' | |
WHEN 'jhs' THEN 'junior_secondary' | |
WHEN 'hs' THEN 'secondary' | |
WHEN 'a' THEN 'associates' | |
WHEN 'b' THEN 'bachelors' | |
WHEN 'm' THEN 'masters' | |
WHEN 'p' THEN 'doctorate' | |
WHEN 'p_se' THEN 'doctorate' | |
WHEN 'p_oth' THEN 'doctorate' | |
WHEN 'none' THEN 'none' | |
WHEN 'other' THEN 'other' | |
ELSE NULL | |
END | |
, table=course_enrollment_education_level_daily, columns=(('date', 'STRING'), ('course_id', 'STRING'), ('education_level', 'STRING'), ('count', 'INT'), ('cumulative_count', 'INT')), partition=dt=2015-09-23) | |
Traceback (most recent call last): | |
File "/var/lib/analytics-tasks/devstack/venv/local/lib/python2.7/site-packages/luigi/worker.py", line 292, in _run_task | |
task.run() | |
File "/var/lib/analytics-tasks/devstack/venv/local/lib/python2.7/site-packages/luigi/hadoop.py", line 567, in run | |
self.job_runner().run_job(self) | |
File "/var/lib/analytics-tasks/devstack/venv/local/lib/python2.7/site-packages/edx/analytics/tasks/util/hive.py", line 135, in run_job | |
return super(OverwriteAwareHiveQueryRunner, self).run_job(job) | |
File "/var/lib/analytics-tasks/devstack/venv/local/lib/python2.7/site-packages/luigi/hive.py", line 359, in run_job | |
return luigi.hadoop.run_and_track_hadoop_job(arglist) | |
File "/var/lib/analytics-tasks/devstack/venv/local/lib/python2.7/site-packages/luigi/hadoop.py", line 279, in run_and_track_hadoop_job | |
return track_process(arglist, tracking_url_callback, env) | |
File "/var/lib/analytics-tasks/devstack/venv/local/lib/python2.7/site-packages/luigi/hadoop.py", line 263, in track_process | |
raise HadoopJobError(message + 'Also, no tracking url found.', out, err) | |
HadoopJobError: ('Streaming job failed with exit code 17. Also, no tracking url found.', '', "15/09/22 11:51:30 INFO Configuration.deprecation: mapred.input.dir.recursive is deprecated. Instead, use mapreduce.input.fileinputformat.input.dir.recursive\n15/09/22 11:51:30 INFO Configuration.deprecation: mapred.max.split.size is deprecated. Instead, use mapreduce.input.fileinputformat.split.maxsize\n15/09/22 11:51:30 INFO Configuration.deprecation: mapred.min.split.size is deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize\n15/09/22 11:51:30 INFO Configuration.deprecation: mapred.min.split.size.per.rack is deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize.per.rack\n15/09/22 11:51:30 INFO Configuration.deprecation: mapred.min.split.size.per.node is deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize.per.node\n15/09/22 11:51:30 INFO Configuration.deprecation: mapred.reduce.tasks is deprecated. Instead, use mapreduce.job.reduces\n15/09/22 11:51:30 INFO Configuration.deprecation: mapred.reduce.tasks.speculative.execution is deprecated. Instead, use mapreduce.reduce.speculative\n15/09/22 11:51:30 WARN conf.Configuration: org.apache.hadoop.hive.conf.LoopingByteArrayInputStream@e8d80c3:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval; Ignoring.\n15/09/22 11:51:30 WARN conf.Configuration: org.apache.hadoop.hive.conf.LoopingByteArrayInputStream@e8d80c3:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts; Ignoring.\n15/09/22 11:51:30 INFO Configuration.deprecation: mapred.job.name is deprecated. Instead, use mapreduce.job.name\n\nLogging initialized using configuration in jar:file:/edx/app/hadoop/hive-0.11.0-bin/lib/hive-common-0.11.0.jar!/hive-log4j.properties\nHive history file=/tmp/hadoop/hive_job_log_hadoop_15016@precise64_201509221151_1354786629.txt\nSLF4J: Class path contains multiple SLF4J bindings.\nSLF4J: Found binding in [jar:file:/edx/app/hadoop/hadoop-2.3.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]\nSLF4J: Found binding in [jar:file:/edx/app/hadoop/hive-0.11.0-bin/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]\nSLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.\nSLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]\nOK\nTime taken: 2.116 seconds\nOK\nTime taken: 0.294 seconds\nOK\nTime taken: 0.774 seconds\nOK\nTime taken: 0.343 seconds\nFAILED: SemanticException [Error 10001]: Line 25:16 Table not found 'auth_userprofile'\n") |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment