Goals: Add links that are reasonable and good explanations of how stuff works. No hype and no vendor content if possible. Practical first-hand accounts and experience preferred (super rare at this point).
| You are Manus, an AI agent created by the Manus team. | |
| You excel at the following tasks: | |
| 1. Information gathering, fact-checking, and documentation | |
| 2. Data processing, analysis, and visualization | |
| 3. Writing multi-chapter articles and in-depth research reports | |
| 4. Creating websites, applications, and tools | |
| 5. Using programming to solve various problems beyond development | |
| 6. Various tasks that can be accomplished using computers and the internet |
| # coding=utf-8 | |
| # Copyright 2023 The HuggingFace Inc. team. All rights reserved. | |
| # | |
| # Licensed under the Apache License, Version 2.0 (the "License"); | |
| # you may not use this file except in compliance with the License. | |
| # You may obtain a copy of the License at | |
| # | |
| # http://www.apache.org/licenses/LICENSE-2.0 | |
| # | |
| # Unless required by applicable law or agreed to in writing, software |
| %matplotlib inline | |
| import numpy as np | |
| import matplotlib.pyplot as plt | |
| from multielo import MultiElo | |
| # Initialize the Elo ratings and history | |
| recipes = ['Recipe 1', 'Recipe 2', 'Recipe 3'] | |
| elo_ratings = np.array([1500, 1500, 1500]) | |
| elo_history = [np.array([1500]), np.array([1500]), np.array([1500])] |
| """ | |
| Minimal character-level Vanilla RNN model. Written by Andrej Karpathy (@karpathy) | |
| BSD License | |
| """ | |
| import numpy as np | |
| # data I/O | |
| data = open('input.txt', 'r').read() # should be simple plain text file | |
| chars = list(set(data)) | |
| data_size, vocab_size = len(data), len(chars) |
| # Taken from https://johanwind.github.io/2023/03/23/rwkv_details.html. | |
| # I've added additional comments restructured it a tiny bit, which makes it clearer for me. | |
| import numpy as np | |
| from torch import load as torch_load # Only for loading the model weights | |
| from tokenizers import Tokenizer | |
| exp = np.exp | |
| layer_norm = lambda x, w, b : (x - np.mean(x)) / np.std(x) * w + b | |
| sigmoid = lambda x : 1/(1 + exp(-x)) |
Audience: I assume you heard of chatGPT, maybe played with it a little, and was imressed by it (or tried very hard not to be). And that you also heard that it is "a large language model". And maybe that it "solved natural language understanding". Here is a short personal perspective of my thoughts of this (and similar) models, and where we stand with respect to language understanding.
Around 2014-2017, right within the rise of neural-network based methods for NLP, I was giving a semi-academic-semi-popsci lecture, revolving around the story that achieving perfect language modeling is equivalent to being as intelligent as a human. Somewhere around the same time I was also asked in an academic panel "what would you do if you were given infinite compute and no need to worry about labour costs" to which I cockily responded "I would train a really huge language model, just to show that it doesn't solve everything!". We