Hey r/SideProjects . I’m currently training my own AI to write as I do. I’m using llama3.1:8B model. Additionally, I’m using AnythingLLM, Vector Database (lanceDB) My tech specs aren’t that great, but they can run the LLM model at a decent pace. I have an Intel i5-12450HX, 16 GB RAM, and RTX 3050 6GB VRAM. I’m training the LLM on my own data, which I’ve collected from various websites, where I’m very active. Instagram: I exported all the DMs I have, only the messages from me, not the other chats. I also exported all my comments on the posts and reels. Telegram: I’m very active here as I have my friend group here, and I have more than 100k messages of myself.How I talk and my personality, too. Discord: Here, where I talk to strangers, is good for data training. Reddit: I’ve exported all my Reddit posts and comments. WhatsApp: Personal chat, and it can give very good insight into my personality. Additionally, I’ve curated a very detailed system prompt for the LLM. I also used a few AI chats to train him on how I ask questions and how I expect a reply from AI. I used the LLMs responses on ZeroGPT, and I’m impressed with the result; it’s only 20~30% AI sounding I’m currently looking for suggestions on how I can improve the training and make it more accurate in replying. Your replies will mean a lot to me. Open to any criticism. Thanks!!! submitted by /u/StealthPhoen1x
Originally posted by u/StealthPhoen1x on r/ArtificialInteligence
