Introducing DoctorGPT! After applying fine-tuning, reinforcement learning, & compilation techniques to Meta’s Llama2 model, I got some amazing results. DoctorGPT is a Large Language Model that can pass the US Medical Licensing Exam. This is an open-source project with a mission to provide everyone their own private doctor. DoctorGPT is a version of Meta’s Llama2 7 billion parameter Large Language Model that was fine-tuned on a Medical Dialogue Dataset, then further improved using Reinforcement Learning & Constitutional AI. Since the model is only 3 Gigabytes in size, it fits on any local device, so there is no need to pay an API to use it. It’s free, made for offline usage which preserves patient confidentiality, and it’s available on iOS, Android, and Web. Pull requests for feature additions and improvements are encouraged. In this video, i’ll go over every single detail of how i built it so that you can do the same yourself. Enjoy!