(by @andrestaltz)
If you prefer to watch video tutorials with live-coding, then check out this series I recorded with the same contents as in this article: Egghead.io - Introduction to Reactive Programming.
using System; | |
using System.Collections.Generic; | |
using System.ComponentModel.DataAnnotations.Schema; | |
using System.Data.Entity; | |
using System.Linq; | |
using System.Linq.Expressions; | |
using System.Reflection; | |
using System.Text; | |
namespace EFExtensions |
(by @andrestaltz)
If you prefer to watch video tutorials with live-coding, then check out this series I recorded with the same contents as in this article: Egghead.io - Introduction to Reactive Programming.
{ | |
"version": "0.1.0", | |
// List of configurations. Add new configurations or edit existing ones. | |
// ONLY "node" and "mono" are supported, change "type" to switch. | |
"configurations": [ | |
{ | |
// Name of configuration; appears in the launch configuration drop down menu. | |
"name": "Run app.js", | |
// Type of configuration. Possible values: "node", "mono". | |
"type": "node", |
# vocabulary = OrderedDict() | |
input_length = None | |
vocabulary_size = max(vocabulary.values()) + 1 | |
weights_w2v = list(map(Word2Vec.__getitem__, vocabulary.keys())) | |
embedding_size len(weights_w2v[0]) | |
nb_classes = 5 | |
# CNN hyperparms | |
nb_filter = 64 | |
filter_length = 5 |
class AttentionLSTM(LSTM): | |
"""LSTM with attention mechanism | |
This is an LSTM incorporating an attention mechanism into its hidden states. | |
Currently, the context vector calculated from the attended vector is fed | |
into the model's internal states, closely following the model by Xu et al. | |
(2016, Sec. 3.1.2), using a soft attention model following | |
Bahdanau et al. (2014). | |
The layer expects two inputs instead of the usual one: |
ChatGPT appeared like an explosion on all my social media timelines in early December 2022. While I keep up with machine learning as an industry, I wasn't focused so much on this particular corner, and all the screenshots seemed like they came out of nowhere. What was this model? How did the chat prompting work? What was the context of OpenAI doing this work and collecting my prompts for training data?
I decided to do a quick investigation. Here's all the information I've found so far. I'm aggregating and synthesizing it as I go, so it's currently changing pretty frequently.