Skip to content

Instantly share code, notes, and snippets.

View anjijava16's full-sized avatar
💭
Awesome

Anjaiah Methuku anjijava16

💭
Awesome
View GitHub Profile
package com.iwinner.runanysql;
import java.io.BufferedReader;
import java.io.FileReader;
import java.io.Reader;
import java.sql.Connection;
import java.sql.DriverManager;
import java.sql.ResultSet;
import java.sql.SQLException;
import java.sql.Statement;
{
insertId: "wr3zn3e373bw"
logName: "projects/iwinner-data-318822/logs/cloudaudit.googleapis.com%2Factivity"
operation: {
first: true
id: "operation-1627046240424-5c7ca39319163-07a56cfa-f16ec9a5"
producer: "compute.googleapis.com"
}
protoPayload: {
@type: "type.googleapis.com/google.cloud.audit.AuditLog"
https://morioh.com/p/4d9c2c2cb9ae?f=5c224490c513a556c9042463&fbclid=IwAR0EA40-1MS3YNSJRkdnGIjAe6of7e3kIRGVRXq1t14aT44eIx4N6ykJ4L8
https://medium.com/@somanathsankaran/window-functions-deep-dive-4aeffebf81e1
https://medium.com/analytics-vidhya/better-spark-logging-and-monitoring-with-spark-listener-c2c5fe86547a
https://medium.com/@somanathsankaran/window-functions-deep-dive-4aeffebf81e1
https://medium.com/analytics-vidhya/better-spark-logging-and-monitoring-with-spark-listener-c2c5fe86547a
Small files spark
https://github.com/capt-blackdron/pyspark_custom_repartitioner/blob/main/custom_repartitioner.py
Scd type 2
https://github.com/capt-blackdron/pyspark-examples/blob/main/SCD_Type_2_Implementation_in_PySpark.ipynb
Spark program
https://stackoverflow.com/questions/59617081/cast-a-column-to-binary-datatype-in-spark
https://stackoverflow.com/questions/53075020/why-does-spark-infer-a-binary-instead-of-an-arraybyte-when-creating-a-datafram
https://stackoverflow.com/questions/41149522/extract-byte-from-spark-binarytype
https://stackoverflow.com/questions/47902691/filtering-by-binary-type-in-spark-sql
https://stackoverflow.com/questions/60627680/spark-binarytype-to-scala-java
SELECT
COUNT(*) TOTAL_QUERIES,
SUM(total_slot_ms/TIMESTAMP_DIFF(end_time,creation_time,MILLISECOND)) AVG_SLOT_USAGE,
SUM(TIMESTAMP_DIFF(end_time,creation_time,SECOND)) TOTAL_DURATION_IN_SECONDS,
AVG(TIMESTAMP_DIFF(end_time,creation_time,SECOND)) AVG_DURATION_IN_SECONDS,
SUM(total_bytes_processed*10e-12) TOTAL_PROCESSED_TB,
EXTRACT (DATE FROM creation_time) AS EXECUTION_DATE,
user_email as USER
FROM `iwinner-data-318822.region-us.INFORMATION_SCHEMA.JOBS_BY_PROJECT`
WHERE state='DONE'
Cloud Spanner has three types of replicas:
i. read-write replicas,
ii. read-only replicas,
iii. witness replicas.
Azure:
ADF/Databricks with Spark- Ingestion framework
ADLS - Data Storage
ADB - Transformations
Data Flows/ Polybase - To load data to Warehouse
Synapse - Datawarehouse
Azure SQL - Metadata Storage
ADF - Orchestration
Logic Apps : Alerts/Email
Azure Devops : Code Deployment
https://www.boredapi.com/api/activity
https://www.mockaroo.com/help/terms_of_use
https://randomuser.me/api/
https://stackoverflow.com/questions/40838036/what-is-the-difference-between-split-by-and-boundary-query-in-sqoop
https://discuss.itversity.com/t/using-boundary-query/18673
https://stackoverflow.com/questions/37206232/sqoop-import-composite-primary-key-and-textual-primary-key