Hi guys, I am a begineer in AI and LLMs. I gained some knowledge and built a RAG based LLM chatbot that uses my PDF to answer. Initially i used ollama to run local Llama 3.2 but I couldn’t get a proper guide on how to host a LLM more over, I have no money to invest as well Later, I changed to Groq API to use the already hosted LLM and managed to get the same output. then, I tried to host it render but it turned to failure cause the storage. I am using Tensor flow, sentnece tranformer embeddings that is occupies more than 500 MB (free tier of render gives only upto 500MB) can any one suggests me any replacement or how to host the my LLM. Or any guidence to run this LLM for free of cost. My aim is just to built and host a chatbot that reads my Q&A pdf and answer based on the pdf. submitted by /u/Royal_Rasengon
Originally posted by u/Royal_Rasengon on r/ArtificialInteligence
