Every time you choose to apply a rule(s), explicitly state the rule(s) in the output. You can abbreviate the rule description to a single word or phrase.
[Brief description ]
- [more description]
- [more description]
- [more description]
| ## Workflow Orchestration | |
| ### 1. Plan Mode Default | |
| - Enter plan mode for ANY non-trivial task (3+ steps or architectural decisions) | |
| - If something goes sideways, STOP and re-plan immediately | |
| - Don't keep pushing. | |
| - Use plan mode for verification steps, not just building | |
| - Write detailed specs upfront to reduce ambiguity | |
| ### 2. Subagent Strategy |
Tip
See the blog post: https://mdinata.my.id/blogs/docker-on-postmarketos/
With postmarketOS and modified kernel (see https://blog.hypriot.com/post/verify-kernel-container-compatibility/), it's possible to natively run Docker on phone. Since postmarketOS runs Alpine Linux natively, there's no need to run VMs, chroot, or whatever it is. Just native. Ain't that cool?
| #!/usr/bin/env sh | |
| # LATEST VERSION OF THIS SCRIPT: https://gist.github.com/swayducky/8ba8f2db156c7f445d562cdc12c0ddb4 | |
| # FORKED FROM: https://gist.github.com/ddwang/0046da801bcb29d241869d37ad719394 | |
| # 1) No longer has a hard-coded COMMIT | |
| # 2) Auto-symlinks a "code" script to avoid wslCode.sh breaking | |
| # HOW TO INSTALL: | |
| # 1) Remove "c:\Users\<USER_NAME>\AppData\Local\Programs\cursor\resources\app\bin" from Windows Environment Settings | |
| # 2) Modify this script with your Windows <USER_NAME> (NOT your WSL username) in the VSCODE_PATH variable |
Disclaimer: I'm not a data scientist or an expert with LLaMA models or LLMs thus I won't cover technical details about LLaMA models, Vicuna or the settings used in my tests. I just wanted to play with LLaMA models and share my setup and results with the community.
I also consider that you have some basic knowledge about using a terminal and running Python scripts.
I will also not cover the installation process of the tools and libraries used in this document but I will provide links to the documentation I used to make this work on my computer.
Finally, I'm not a native English speaker so please excuse my English mistakes 🙃
I wanted to try to run a LLaMA model on my computer. Since I had absolutely no knowledge about this I started by reading a lot of documentation and articles on the Internet.
Create a fresh new Next js project. https://nextjs.org/docs/getting-started#automatic-setup
pnpm create next-app --typescript
install prettier
| const ALGOLIA_APP_ID = "<your algolia app id>"; | |
| addEventListener('fetch', event => { | |
| try { | |
| const request = event.request; | |
| if (request.method.toUpperCase() === 'POST') return event.respondWith(handlePostRequest(event)); | |
| return handleRequest(request); | |
| } catch (e) { | |
| return event.respondWith(new Response('Error thrown ' + e.message)); | |
| } |