Skip to content

Instantly share code, notes, and snippets.

View Databracket9's full-sized avatar
🎯
Maximum Effort

JayReddy Databracket9

🎯
Maximum Effort
View GitHub Profile
@Databracket9
Databracket9 / API to Postgres ETL
Created July 7, 2024 09:51
End-to-End Data Engineering with Pandas
import pandas as pd
import requests
import configparser
import psycopg2
# raw_data = requests.get("https://storage.googleapis.com/generall-shared-data/startups_demo.json")
# with open("raw_files/raw_data.json", "w") as f:
# f.write(raw_data.text)
@Databracket9
Databracket9 / snowpark_load.py
Created May 15, 2024 14:06
Demonstration to learn how to perform ETL operation on snowflake datasets using Python with Snowpark and store in S3.
import snowflake.snowpark as snowpark
from datetime import date
def create_stage(session: snowpark.Session, stage_name: str, s3_arn:str, aws_id:str, aws_secret:str):
query = "create or replace temporary stage {} url='{}' CREDENTIALS=(AWS_KEY_ID='{}' AWS_SECRET_KEY='{}')".format(stage_name, s3_arn, aws_id, aws_secret)
state = snowpark.Session.sql(self=session, query=query)
return state
def query_data(session: snowpark.Session, query: str, date: str):
table = snowpark.Session.sql(self=session, query=query)
@Databracket9
Databracket9 / duckdb_postgres_pyspark_tutorial
Created May 9, 2024 19:05
Data Engineering with DuckDb Tutorial
#library import
import duckdb
import configparser
from duckdb.experimental.spark.sql import SparkSession as session
from duckdb.experimental.spark.sql.functions import col, when, lit
#read configs from secrets file
config = configparser.ConfigParser()
config.read('duck_db_demo/secrets.ini')
user = config['POSTGRES']['USER']