Skip to content

Instantly share code, notes, and snippets.

@FanaHOVA
Created August 9, 2023 02:59
Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save FanaHOVA/0b52f5c152330fe6c644cb4fa020441b to your computer and use it in GitHub Desktop.
Save FanaHOVA/0b52f5c152330fe6c644cb4fa020441b to your computer and use it in GitHub Desktop.

Chapters:

  • [00:00:37] Introduction to Tianqi Chen
  • [00:01:07] Background on creating XGBoost
  • [00:04:11] Motivation and process behind creating XGBoost
  • [00:06:53] Comparing tree-based models vs deep learning
  • [00:12:21] Overview of how TVM and ONNX work together
  • [00:16:23] Background and motivation behind creating TVM
  • [00:18:30] Importance of machine learning systems research
  • [00:20:02] Overview of MLC and running models on phones
  • [00:24:43] Deploying models compiled with TVM in different languages
  • [00:29:05] Optimizations done in TVM like kernel fusion and quantization
  • [00:31:57] Transformations done in TVM to optimize performance
  • [00:33:19] Using int4 quantization for inference of language models
  • [00:36:45] Comparison of MLC to other model optimization projects
  • [00:39:54] Running large language models like GPT-3 in the browser with WebLM
  • [00:42:49] Integrating browser models into applications
  • [00:47:35] Motivation behind building OctoAI model serving
  • [00:50:43] Challenges in abstracting model serving from end users
  • [00:53:49] Hopes for an open ecosystem of diverse AI models interacting
  • [00:55:52] Surprise at the emergence of conversational AI
  • [00:57:20] Interest in continuous and lifelong learning for AI

GPT-3.5 16k title suggestions:

  1. "From RLHF to RLHB: The Case for Learning from Human Behavior"
  2. "Commoditizing the Petaflop"
  3. "Llama 2: The New Open LLM SOTA"
  4. "FlashAttention 2: making Transformers 800% faster w/o approximation"
  5. "Mapping the future of truly Open Models and Training Dolly for $30"
  6. "The Evolution of Machine Learning Compilation and Universal Deployment"
  7. "Optimizing Model Runtimes: The OctoAI Approach"
  8. "The Future of AI: Continuous Learning and Lifelong Adaptation"

Claude's title suggestions: Here are 8 suggested title options for the podcast transcript:

  1. "The Evolution of Machine Learning Compilation with Tianqi Chen"

  2. "From XGBoost to TVM: Tianqi Chen on Systems for ML"

  3. "Tianqi Chen on Building Systems and Tools for the ML Community"

  4. "Inside the Mind of a Machine Learning Systems Innovator"

  5. "Democratizing Machine Learning with MLC and WebLM"

  6. "Tianqi Chen on Compiling Models for Universal Deployment"

  7. "Bringing Large Language Models to the Browser and Beyond"

  8. "TQ Chen on His Journey from XGBoost to TVM and MLC"

GPT-3.5 16k tweet suggestions:

  1. "In our latest podcast episode, we dive into the world of ML computer science with Tianqi Chen, the creator of Apache TVM, XGBoost, and MXNet. Get ready for a fascinating discussion on the evolution of machine learning models and the power of open source projects. #ML #OpenSource"

  2. "TQ Chen, the mastermind behind Apache TVM, XGBoost, and MXNet, shares his personal side and his passion for sketching out high-level architectures in our latest podcast episode. Tune in to discover the creative process behind building groundbreaking ML frameworks. #ML #Innovation"

  3. "Did you know that XGBoost was originally created to test the hypothesis that tree-based models could match the performance of deep learning networks? Learn more about the fascinating journey of XGBoost and its creator Tianqi Chen in our latest podcast episode. #XGBoost #DeepLearning"

  4. "Discover the future of machine learning compilation with Tianqi Chen, the co-founder of OctoML, in our latest podcast episode. Learn about the exciting advancements in ML systems and how they're reshaping the landscape of AI. #MLSystems #AI"

  5. "Exciting news! OctoML has released OctoAI, a revolutionary compute service that optimizes model runtimes and accelerates compilation. Join us as we explore the world of AI optimization and the power of OctoAI in our latest podcast episode. #AI #Optimization"

  6. "Get ready to unlock the full potential of AI with OctoML's OctoAI. Discover how this cutting-edge compute service is revolutionizing model runtimes and accelerating compilation. Tune in to our latest podcast episode to learn more. #AI #OctoAI"

  7. "Curious about the future of AI and ML? Join us as we chat with TQ Chen, the creator of Apache TVM, XGBoost, and MXNet, in our latest podcast episode. Gain insights into the evolution of ML frameworks and the exciting possibilities that lie ahead. #AI #ML"

  8. "From sketching high-level architectures to building groundbreaking ML frameworks, TQ Chen has done it all. Don't miss our latest podcast episode, where we dive into the fascinating world of ML computer science with the man behind Apache TVM, XGBoost, and MXNet. #ML #Tech"

Claude's tweet suggestions: Here are 8 suggested tweets to share the podcast episode:

  1. Learn how @tqchen built systems like XGBoost, Apache TVM, and OctoML to optimize ML models. His goal: enable universal deployment from cloud to edge. 🚀

  2. Key takeaways:

  • System optimization is crucial for delivering performant ML applications
  • Compilation tech like TVM allows models to run on any hardware
  • Open ecosystem for diverse AI agents that can personalize and collaborate
  1. "When you build AI applications, it takes algorithm, system optimization, and data curation." - @tqchen

  2. @tqchen started with hacking CUDA kernels by hand. Now his compiler TVM auto-generates optimized code for any hardware.

  3. Running big models on phones? @tqchen compiled Lamda2 to ship in the new MLC app. His goal: universal deployment from cloud to edge.

  4. Browser-based AI? @tqchen compiled a 70B parameter model to run smoothly on WebGPU thanks to TVM and WebLM. The future is here.

  5. Key ideas from @tqchen:

  • Tree models still shine for tabular data vs DNNs
  • Systems research enables breakthroughs in ML productivity
  • Open ecosystem for personalized, collaborative AI agents
  1. "Being able to build open source and work with communities is really rewarding... See how research translates to real impact." - @tqchen
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment