How are you all? What’cha doing right now?
Hi. Testing a LoRA finetune AI model for someone. U?
Being behind on an automation cleanup script. I didn’t do very well mentally 2 days ago and had to take a day off, I do feel much better now though.
What is automation cleanup? Like garbage collection for memory stuff?
Clean up of orphaned dynamic cloud computing resources once they are no longer in use, should the main pipeline fail to do so, such as due to issues like the Jenkins server crashing, or from bugs from the base machine creation pipeline failing to delete them.
Having VMs continue to run without anyone’s awareness for however many hours it takes someone to go in and manually clean them up can be very expensive, very very expensive. Will pay for my salary in wasted computing cost expensive. I really should get back to work, see ya 👋.
Hey j4k3, just completed 1 half—the easier half—of the automation. Hope I can get the rest up soon too.
You got this.
Don’t be afraid to rewrite too! That is what gets me with code. For some stupid reason I do this with CAD naturally but with scripts and code I battle way too long and often give up when I probably should be iterating from scratch as my logic and layers get messy. I lack the overview perspective in code that I have with CAD. Like I know the topological naming issue really well in CAD. The truncation of π and the temporal linearity of math relationship between coordinates in the tree are the TNI. In code my understanding of schedulers is likely my biggest weak point wall, but also advanced algorithms.
I actually initially got into AI hoping to help with custom tailored learning to help follow the 2nd-3rd year CS curriculum on my own.
Stick with it! You’ll get there! GL!
Urgh, I tried the AI stuff, and then I had to begin manually rewriting code. I feel like a PR reviewer trying to understand what some stranger has written. It’s surprisingly far less fun then doing it from scratch.
And thanks for the encouragement, I really do appreciate it.
I’m probably not doing anything at the same complexity when I play with AI and code. I’ve been playing with this stuff singe around July of 2023 when I got offline open weights capable hardware to run my own models. I dove into it with both feet in all spaces. Now, I know how to talk to it effectively. Prompting language is very different than regular human communication. There is a ton of nuance and unintuitive aspects that most people fail to understand. Even foundation model researchers do not have a solid grasp on how models think internally. Like I see stuff that is obvious to me in mistakes they are making all the time. For instance, name-1 (User) and name-2 (Not) are not distinct entities. That is just a naming convention to help keep the json straight for the human creating model loader code. The model has no identity or sense of self at all. It simply continues the text by assuming a profile for the place where the total context leaves off. This is why they often continue asking their own Q&A after answering your question. Getting the model to stop doing this is totally arbitrary and hard. The reason it happens is because everything with a LLM is roleplaying, even acting like a coding assistant.
One major fundamental aspect people fail to understand in this paradigm is that every character present in the context has an assumed character profile; yes, even the code assistant. You see, these profiles are made up as soon as any character is implied in the context. If you explicitly define parts of this profile for any character, it will override the assumptions and redactor the character. However, you will not override/eliminate any earlier defined aspects in the total context. There are a ton of hidden steganographic tokens that tend to hold secret meanings to a model. I don’t think anyone in academia has explored this but I know about several that alter model alignment considerably when banned during inference (not through model loader code). So yeah don’t trust long contexts with any previous information when accuracy matters. However, you need a ton of context to define profiles well for all characters, or you need extremely concise and precision ways of defining your characters.
For instance, I name my assistant “Richard Stallman’s AI Assistant:” when I am doing anything in bash or Linux. Stallman was originally an AI researcher. Supposedly some of his work is still in the chain of research used in current models. His online presence and what his history implies about the expected competency of an AI model is super powerful.
Sentex (YT) – guy literally wrote the book on AI stuff – had a great suggestion awhile back. Just take your entire block of code, paste it into a leading model and tell it to make a tutorial about how the code works.
It is complicated stuff. I have simply had the time to play with AI due to disability and my situation in life. Understanding alignment thinking is key to understanding the limits of AI and how to push it much further than most people. That is a deep rabbit hole. If you want to get into it emacs has gptel. There is no replacement for running larger models like a 70b+ on your own hardware. Oobabooga is a good starting point if you learn the notepad. That runs a llama.cpp server front end. Llama.cpp is what most people are running in the background. It is better than stuff like ollama because it can split the model between CPU and GPU while using system memory and that allows you to run much larger quantized models like a 70B with a 16GB GPU and 64GB+ of system memory. The proprietary online inference models are a little better but not by much. The key is that you will have nothing holding you back from inferencing anything and exploring just because you can. I would not have tried thousands of things if I had been paying for inference tokens.
Anyways, I don’t fault you if it is not for you. Don’t let it distract you from getting important stuff done!
Oh yeah, I’m definitely interested in exploring the local AI stuff. They’re not approved for use at my company however, so anything they generate can’t be committed, and anything you give the AI can’t contain source code, for liability purposes.
My options at work right now are GitHub Copilot and Cursor. I can’t use anything else really, without going through a very pain process of approvals, or unless I’ll like to be looking for a new job very soon.