Skip to content

Instantly share code, notes, and snippets.

View paulnicholsen27's full-sized avatar

Paul Nicholsen paulnicholsen27

View GitHub Profile
25/10/10 13:19:32 ERROR Utils: Aborting task
scala.MatchError: java.lang.NoClassDefFoundError: Could not initialize class org.xerial.snappy.Snappy (of class java.lang.NoClassDefFoundError)
at org.apache.spark.sql.execution.datasources.v2.FileDataSourceV2$.attachFilePath(FileDataSourceV2.scala:127)
at org.apache.spark.sql.execution.datasources.FileScanRDD$$anon$1.hasNext(FileScanRDD.scala:142)
at org.apache.spark.sql.execution.FileSourceScanExec$$anon$1.hasNext(DataSourceScanExec.scala:695)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage1.columnartorow_nextBatch_0$(Unknown Source)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage1.processNext(Unknown Source)
at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
at org.apache.spark.sql.execution.WholeStageCodegenEvaluatorFactory$WholeStageCodegenPartitionEvaluator$$anon$1.hasNext(Who
Writing DataFrame to Parquet file: /data/users/nicholsenpm/airflow_extractions/BTRIS_CB_Color_Coding_09152025_115611/output/PRES_subjects_processed/PRES_subjects_processed.parquet
25/09/15 13:24:06 WARN MemoryManager: Total allocation exceeds 95.00% (1,020,054,720 bytes) of heap memory
Scaling row group sizes to 95.00% for 8 writers
25/09/15 13:24:06 ERROR Utils: Aborting task (0 + 8) / 8]
org.apache.spark.SparkUpgradeException: [INCONSISTENT_BEHAVIOR_CROSS_VERSION.WRITE_ANCIENT_DATETIME] You may get a different result due to the upgrading to Spark >= 3.0:
writing dates before 1582-10-15 or timestamps before 1900-01-01T00:00:00Z
into Parquet files can be dangerous, as the files may be read by Spark 2.x
or legacy versions of Hive later, which uses a legacy hybrid calendar that
is different from Spark 3.0+'s Proleptic Gregorian calendar. See more
25/09/10 10:02:45 INFO FileSourceStrategy: Post-Scan Filters:
25/09/10 10:02:45 INFO ShufflePartitionsUtil: For shuffle(6), advisory target size: 67108864, actual target size 1048576, minimum partition size: 1048576
25/09/10 10:02:45 INFO SparkContext: Starting job: $anonfun$withThreadLocalCaptured$1 at FutureTask.java:264
25/09/10 10:02:45 INFO DAGScheduler: Got job 16 ($anonfun$withThreadLocalCaptured$1 at FutureTask.java:264) with 1 output partitions
25/09/10 10:02:45 INFO DAGScheduler: Final stage: ResultStage 26 ($anonfun$withThreadLocalCaptured$1 at FutureTask.java:264)
25/09/10 10:02:45 INFO DAGScheduler: Parents of final stage: List(ShuffleMapStage 25)
25/09/10 10:02:45 INFO DAGScheduler: Missing parents: List()
25/09/10 10:02:45 INFO DAGScheduler: Submitting ResultStage 26 (MapPartitionsRDD[65] at $anonfun$withThreadLocalCaptured$1 at FutureTask.java:264), which has no missing parents
25/09/10 10:02:45 INFO MemoryStore: Block broadcast_29 stored as values in memory (estimated size 113.4 KiB, free 42
(pipeline_extractor) nicholsenpm@ncias-d3613-v /data/users/nicholsenpm $ /opt/pipelines/spark-3.5.6-bin-hadoop3-scala2.13/bin/spark-submit --conf spark.sql.parquet.compression.codec=gzip airflow_extractions/BTRIS_CB_Color_Coding_08212025_140528/src/btris_procedure_processed/btris_procedure_processed.jar airflow_extractions/BTRIS_CB_Color_Coding_08212025_140528/output/btris_observation_general_clean airflow_extractions/BTRIS_CB_Color_Coding_08212025_140528/output/btris_red_ancestor_descendant_clean airflow_extractions/BTRIS_CB_Color_Coding_08182025_133543/airflow_extractions/BTRIS_CB_Color_Coding_08212025_140528/output/btris_procedure_processed
25/08/26 11:49:51 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
25/08/26 11:49:51 INFO SparkContext: Running Spark version 3.5.6
25/08/26 11:49:51 INFO SparkContext: OS info Linux, 4.18.0-553.69.1.el8_10.x86_64, amd64
25/08/26 11:49:51 INFO SparkContext: Java version 11.0.27
25/08/26 11:49:5
import json
import csv
project_folder = "airflow_extractions/BTRIS_CB_Color_Coding_08192025_100349/"
path_to_action_steps = f"{project_folder}action_steps.json"
with open(path_to_action_steps, "r") as f:
data = json.load(f)
(function (exports) {
'use strict';
var instagramCss = "/* Instagram feed, stories, recommended accounts */\nhtml:not([data-nfe-enabled='false']) main > :not(#nfe-container) {\n\tdisplay: none;\n}\n\nhtml:not([data-nfe-enabled='false']) main > #nfe-container {\n\twidth: 100%;\n\tfont-size: 24px;\n\tpadding: 128px;\n}\n";
var twitterCss = "/* Twitter */\nhtml:not([data-nfe-enabled='false'])\n\tdiv[data-testid='primaryColumn']\n\t> div:last-child\n\t> div:nth-child(4)\n\t> #nfe-container {\n\tpadding: 16px;\n}\n\nhtml:not([data-nfe-enabled='false']) div[aria-label*='Timeline'],\nhtml:not([data-nfe-enabled='false']) div[data-testid=\"primaryColumn\"] > div:last-child > div:last-child\n{\n\topacity: 0 !important;\n\tpointer-events: none !important;\n}\n\n/* \"What's Happening\" section on Twitter */\n[data-testid='sidebarColumn'] [role='region'] {\n\topacity: 0 !important;\n\tpointer-events: none !important;\n\theight: 0 !important;\n}\n";
var linkedinCss = "/* LinkedIn */\nhtml:not([data-nfe-enabled='fals
# Array Practice
## Practice with Integers
Integers - an array from 0 to 100:
```
(0..100).to_a
```
require 'bundler/setup'
Bundler.require
require_rel '../app'
describe "Challenge" do
before(:all) do
@customer1 = Customer.new("Bob", "Bobbington")
@customer2 = Customer.new("Mary", "Maryberg")
@customer3 = Customer.new("Bob", "SomeoneElse")
## The names of those who have gone before you
| Cohort | Name |
| ------------------ | ------------------------------- |
| DC Web 082718 | Puts rspec on my :name
| Houston Web 082718 | Boolean Icing (but the chant to exit feelings is "Fork Repos Git Money"; inside-joke stuff)
| Dumbo Web 080618 | Kiss My Hash
| NY Web 080618 | RSpec the Wealth
| DC Web 071618 | Good Migrations
require 'bundler/setup'
Bundler.require
require_rel '../app'
describe "Challenge" do
before(:all) do
@customer1 = Customer.new("Bob", "Bobbington")
@customer2 = Customer.new("Mary", "Maryberg")
@customer3 = Customer.new("Bob", "SomeoneElse")