✷ This project does not touch the code in the clearml_serving directory. However, due to its awesomeness and tight integration with this demo, it was pertinant to bring along for the ride. Enjoy!
Um. Ok. Well. This is a step by step guide to building and deploying your own Custom GPT and API with End to End ClearML integration. Fun stuff.
Embark on a comprehensive journey to create, deploy, and leverage your very own Custom GPT and API, fully integrated with ClearML for a seamless experience.
This step-by-step tutorial takes you through the entire process, starting with the construction of a robust PyTorch model that combines the prowess of a BERT/BiLSTM
with the interpretability of an Explainable Boosting Machine
(EBM) from InterpretML.
Discover how to bring your API to life using the power of HuggingFace
, PyTorch
, ClearML
, ClearML-Serving
, Docker
, NVIDIA Triton
, and MORE ensuring your model is not just a marvel of machine learning but also a fully operational service. Dive into the world of custom GPT actions, designing a system that communicates effectively with your API, and explore the innovative concept of Retrieval Augmented Generation
(RAG) to enhance your GPT's responses with real-time, llm based enhanced explainability.
Whether you're a seasoned data scientist or an enthusiastic beginner ✌, this guide promises to equip you with the knowledge and tools needed to bring your AI visions to reality.
Wow, right? I hope it's that exciting. I hope I do all that.
✌ That last part may be off. This is not easy stuff. You will work at getting it all going if you wish to recreate it. But just dive into the areas that are new to you, and kick its ass.
😻 → Build API
😂 → Have a Laugh
🤟 If your organization is looking for someone with my skillset, I am available as of 2-12-2024. Send me a message or check out my resume at http://www.mikelange.com I look forward to chatting!
- → Hugging Face: https://huggingface.co/
- → Interpret ML: https://interpret.ml/
- → Apache https://www.apache.org/
- → ClearML: https://clear.ml/
- → Thanks
Victor
For the awesome Vids! - → Kaggle: https://Kaggle.com
- → PyTorch: https://PyTorch.org
- → Tensorflow: https://www.tensorflow.org/
- → OpenAI: https://openai.com
- → Docker: https://docker.com
- → NVIDIA: https://nvidia.com
- → Optuna: https://optuna.org/
- → Anaconda: https://anaconda.org/
- → Spacy: https://spacy.io/
- → Berkley Neural Parser: https://spacy.io/universe/project/self-attentive-parser
- → scikit-learn https://scikit-learn.org
- → BERT: https://huggingface.co/docs/transformers/en/model_doc/bert
- → Empath: https://github.com/Ejhfast/empath-client
- → textstat: https://pypi.org/project/textstat/
- and all the others I didn't mention.
TODOS:
- Improve the models. Help needed.
- Clean up code.
- ClearML Dashboards
- ClearML - Improve Integration
- Automate Training
- Improved metrics with training and inference
- Grafana dashboards
- Configure helm charts for k8 deployment
- Speed up feature gen pipeline
- Build EBM to ONYX pipeline to serve full ensemble on Triton (https://github.com/interpretml/ebm2onnx)
-
Mad-Lib
the training data gen prompts to do GAN optimizations. (explain this process better) - Greatly reduce knoledge file
- impliment https://github.com/interpretml/TalkToEBM and
- https://github.com/interpretml/interpret-text to expand explability.
References: