Skip to content

Instantly share code, notes, and snippets.

View tonidy's full-sized avatar

toni dy tonidy

View GitHub Profile
@ScinDBad
ScinDBad / IDX_Jupyter_guide.md
Last active October 9, 2024 15:30
A short guide to start python projects and jupyter notebooks in Project IDX by Google. By Adrián Vinueza

Guide: Developing Jupyter Notebooks in Google IDX

What is Google IDX?

Google IDX (Integrated Development Experience) is a cloud-based integrated development environment (IDE) provided by Google, designed to simplify the setup, management, and collaboration of projects entirely in the cloud. IDX shows promising potential to establish itself over time, offering an intriguing cloud alternative similar to Visual Studio Code or Anaconda.

Creating Jupyter Notebooks or Python Projects in Google IDX

Jupyter Notebooks are widely used for interactive computing and data analysis, ideal for tasks ranging from data exploration to developing machine learning models. Setting up your environment in Google IDX for Jupyter Notebooks or Python projects is straightforward and provides the advantage of cloud-based programming.

  1. Go to https://idx.dev/ and get started.
@guest271314
guest271314 / harmony.md
Last active February 10, 2025 16:27
Why I use node, deno, bun, qjs, tjs at the same time

Why I use node, deno, bun, qjs, tjs at the same time.

Winds up being a (not the) rather comprehensive JavaScript toolbox. The idea being for the modern JavaScript programmer can use all of the tools available for a given requirement, task or job, without preference for any. No external bundlers or compilers are needed. No frameworks are needed. I can use qjs or tjs for systems with minimal RAM and disk space; and when I want to use Web API's deno makes an effort to provide those interfaces. In some cases I can run the exact same code in bun, deno, and node, which provides a means to perform 1:1 testing as to performance.

There's probably a few things I am unintentionally omitting below. These are just a brief synposis. I'll update accordingly.

@guest271314
guest271314 / javascript_engines_and_runtimes.md
Last active April 30, 2025 01:29
A list of JavaScript engines, runtimes, interpreters

V8 is Google’s open source high-performance JavaScript and WebAssembly engine, written in C++. It is used in Chrome and in Node.js, among others. It implements ECMAScript and WebAssembly, and runs on Windows 7 or later, macOS 10.12+, and Linux systems that use x64, IA-32, ARM, or MIPS processors. V8 can run standalone, or can be embedded into any C++ application.

SpiderMonkey is Mozilla’s JavaScript and WebAssembly Engine, used in Firefox, Servo and various other projects. It is written in C++, Rust and JavaScript. You can embed it into C++ and Rust projects, and it can be run as a stand-alone shell. It can also be [compiled](https://bytecodealliance.org/articles/making-javascript-run-fast-on

@scivision
scivision / CMakeLists.txt
Last active March 14, 2025 19:07
OpenMP with CMake
cmake_minimum_required(VERSION 3.19)
project(OpenMPdemo LANGUAGES C)
find_package(OpenMP COMPONENTS C REQUIRED)
add_executable(hello hello_openmp.c)
target_link_libraries(hello PRIVATE OpenMP::OpenMP_C)
@ShaneShipston
ShaneShipston / php.ini
Created July 23, 2023 16:20
NativePHP Windows Support
[PHP]
;;;;;;;;;;;;;;;;;;;
; About php.ini ;
;;;;;;;;;;;;;;;;;;;
; PHP's initialization file, generally called php.ini, is responsible for
; configuring many of the aspects of PHP's behavior.
; PHP attempts to find and load this configuration from a number of locations.
; The following is a summary of its search order:
@bvisness
bvisness / wasm_in_ci.md
Last active December 21, 2024 19:33
A quick guide to running wasm files outside the browser

General info

The following code should work to test WebAssembly in various JS runtimes with minor modifications:

const bytes = /* read a .wasm file somehow */;
const mod = new WebAssembly.Module(bytes);
const instance = new WebAssembly.Instance(mod, { /* imports */ });

const { foo, bar } = instance.exports;
alias nvim-lazy="NVIM_APPNAME=LazyVim nvim"
alias nvim-chad="NVIM_APPNAME=NvChad nvim"
alias nvim-astro="NVIM_APPNAME=AstroNvim nvim"
function nvims() {
items=("default" "LazyVim" "NvChad" "AstroNvim")
config=$(printf "%s\n" "${items[@]}" | fzf --prompt=" Neovim Config  " --height=~50% --layout=reverse --border --exit-0)
if [[ -z $config ]]; then
echo "Nothing selected"
return 0
@gekigek99
gekigek99 / cpu usage percent.go
Last active September 1, 2024 12:35
calculate the average cpu usage percent of the last second of a given pid
// This script calculates the average cpu usage percent of the last second (of a given pid, in this case the script itself).
// (*process.Process).CPUPercent() from "github.com/shirou/gopsutil/process" returns the average cpu usage percent since the process was started.
// If you call cpuPercent(p) every 10 seconds then the average cpu usage percent of the last 10 seconds is returned.
//
// cpuPercent first call returns the average cpu percent usage since the start of the process.
//
// Remember that if the process you are analyzing changes pid you have to update `pTracker` with the new pid (+new info) and remove the old pid.
// (not the case for this example though)
package main
@Stroniax
Stroniax / SingleCollectionBenchmark.cs
Created February 16, 2023 19:45
SingleCollection Benchmark
using BenchmarkDotNet.Attributes;
using BenchmarkDotNet.Running;
using System.Collections;
using System.Runtime.CompilerServices;
using System.Runtime.InteropServices;
namespace SingleCollectionBenchmark;
/// <summary>
/// <code>
@devinschumacher
devinschumacher / cloud-gpus.md
Last active April 28, 2025 16:42
Cloud GPUs // The Best Servers, Services & Providers [RANKED!]

Cloud GPUs: Servers, Providers & Everything You Would Ever Need

Your company's GPU computing strategy is essential whether you engage in 3D visualization, machine learning, AI, or any other form of intensive computing.

There was a time when businesses had to wait for long periods of time while deep learning models were being trained and processed. Because it was time-consuming, costly, and created space and organization problems, it reduced their output.

This problem has been resolved in the most recent GPU designs. Because of their high parallel processing efficiency, they are well-suited for handling large calculations and speeding up the training of your AI models.

When it comes to deep learning, good Cloud GPUs can speed up the training of neural networks by a factor of 250 compared to CPUs, and the latest generation of cloud GPUs is reshaping data science and other emerging technologies by delivering even greater performance