Skip to content

Instantly share code, notes, and snippets.

@natea
Created September 22, 2015 11:54
Show Gist options
  • Save natea/4dde0f50112ec2508a30 to your computer and use it in GitHub Desktop.
Save natea/4dde0f50112ec2508a30 to your computer and use it in GitHub Desktop.
2015-09-22 11:51:34,841 INFO 12850 [luigi-interface] hadoop.py:234 - FAILED: SemanticException [Error 10001]: Line 25:16 Table not found 'auth_userprofile'
2015-09-22 11:51:34,980 ERROR 12850 [luigi-interface] worker.py:304 - [pid 12850] Worker Worker(salt=039843096, host=precise64, username=hadoop, pid=12850) failed HiveTableFromParameterQueryTask(overwrite=False, warehouse_path=hdfs://localhost:9000/edx-analytics-pipeline/warehouse/, insert_query=
SELECT
ce.date,
ce.course_id,
CASE p.level_of_education
WHEN 'el' THEN 'primary'
WHEN 'jhs' THEN 'junior_secondary'
WHEN 'hs' THEN 'secondary'
WHEN 'a' THEN 'associates'
WHEN 'b' THEN 'bachelors'
WHEN 'm' THEN 'masters'
WHEN 'p' THEN 'doctorate'
WHEN 'p_se' THEN 'doctorate'
WHEN 'p_oth' THEN 'doctorate'
WHEN 'none' THEN 'none'
WHEN 'other' THEN 'other'
ELSE NULL
END,
SUM(ce.at_end),
COUNT(ce.user_id)
FROM course_enrollment ce
LEFT OUTER JOIN auth_userprofile p ON p.user_id = ce.user_id
GROUP BY
ce.date,
ce.course_id,
CASE p.level_of_education
WHEN 'el' THEN 'primary'
WHEN 'jhs' THEN 'junior_secondary'
WHEN 'hs' THEN 'secondary'
WHEN 'a' THEN 'associates'
WHEN 'b' THEN 'bachelors'
WHEN 'm' THEN 'masters'
WHEN 'p' THEN 'doctorate'
WHEN 'p_se' THEN 'doctorate'
WHEN 'p_oth' THEN 'doctorate'
WHEN 'none' THEN 'none'
WHEN 'other' THEN 'other'
ELSE NULL
END
, table=course_enrollment_education_level_daily, columns=(('date', 'STRING'), ('course_id', 'STRING'), ('education_level', 'STRING'), ('count', 'INT'), ('cumulative_count', 'INT')), partition=dt=2015-09-23)
Traceback (most recent call last):
File "/var/lib/analytics-tasks/devstack/venv/local/lib/python2.7/site-packages/luigi/worker.py", line 292, in _run_task
task.run()
File "/var/lib/analytics-tasks/devstack/venv/local/lib/python2.7/site-packages/luigi/hadoop.py", line 567, in run
self.job_runner().run_job(self)
File "/var/lib/analytics-tasks/devstack/venv/local/lib/python2.7/site-packages/edx/analytics/tasks/util/hive.py", line 135, in run_job
return super(OverwriteAwareHiveQueryRunner, self).run_job(job)
File "/var/lib/analytics-tasks/devstack/venv/local/lib/python2.7/site-packages/luigi/hive.py", line 359, in run_job
return luigi.hadoop.run_and_track_hadoop_job(arglist)
File "/var/lib/analytics-tasks/devstack/venv/local/lib/python2.7/site-packages/luigi/hadoop.py", line 279, in run_and_track_hadoop_job
return track_process(arglist, tracking_url_callback, env)
File "/var/lib/analytics-tasks/devstack/venv/local/lib/python2.7/site-packages/luigi/hadoop.py", line 263, in track_process
raise HadoopJobError(message + 'Also, no tracking url found.', out, err)
HadoopJobError: ('Streaming job failed with exit code 17. Also, no tracking url found.', '', "15/09/22 11:51:30 INFO Configuration.deprecation: mapred.input.dir.recursive is deprecated. Instead, use mapreduce.input.fileinputformat.input.dir.recursive\n15/09/22 11:51:30 INFO Configuration.deprecation: mapred.max.split.size is deprecated. Instead, use mapreduce.input.fileinputformat.split.maxsize\n15/09/22 11:51:30 INFO Configuration.deprecation: mapred.min.split.size is deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize\n15/09/22 11:51:30 INFO Configuration.deprecation: mapred.min.split.size.per.rack is deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize.per.rack\n15/09/22 11:51:30 INFO Configuration.deprecation: mapred.min.split.size.per.node is deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize.per.node\n15/09/22 11:51:30 INFO Configuration.deprecation: mapred.reduce.tasks is deprecated. Instead, use mapreduce.job.reduces\n15/09/22 11:51:30 INFO Configuration.deprecation: mapred.reduce.tasks.speculative.execution is deprecated. Instead, use mapreduce.reduce.speculative\n15/09/22 11:51:30 WARN conf.Configuration: org.apache.hadoop.hive.conf.LoopingByteArrayInputStream@e8d80c3:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval; Ignoring.\n15/09/22 11:51:30 WARN conf.Configuration: org.apache.hadoop.hive.conf.LoopingByteArrayInputStream@e8d80c3:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts; Ignoring.\n15/09/22 11:51:30 INFO Configuration.deprecation: mapred.job.name is deprecated. Instead, use mapreduce.job.name\n\nLogging initialized using configuration in jar:file:/edx/app/hadoop/hive-0.11.0-bin/lib/hive-common-0.11.0.jar!/hive-log4j.properties\nHive history file=/tmp/hadoop/hive_job_log_hadoop_15016@precise64_201509221151_1354786629.txt\nSLF4J: Class path contains multiple SLF4J bindings.\nSLF4J: Found binding in [jar:file:/edx/app/hadoop/hadoop-2.3.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]\nSLF4J: Found binding in [jar:file:/edx/app/hadoop/hive-0.11.0-bin/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]\nSLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.\nSLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]\nOK\nTime taken: 2.116 seconds\nOK\nTime taken: 0.294 seconds\nOK\nTime taken: 0.774 seconds\nOK\nTime taken: 0.343 seconds\nFAILED: SemanticException [Error 10001]: Line 25:16 Table not found 'auth_userprofile'\n")
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment