OpenCoder: The OPEN Cookbook For Top-Tier Code LLMs
            Welcome to the last arXiv Dive of 2024! Every other week we have been diving into interesting research papers in
        
     
    
            
    
        
    
        LLaVA-CoT: Let Vision Language Models Reason Step-By-Step
            When it comes to large language models, it is still the early innings. Many of them still hallucinate, fail to
        
     
    
            
    
        
    
        How Upcycling MoEs Beat Dense LLMs
            In this Arxiv Dive, Nvidia researcher, Ethan He, presents his co-authored work Upcycling LLMs in Mixture of Experts (MoE). He
        
     
    
            
    
        
    
        Thinking LLMs: General Instruction Following with Thought Generation
            The release of OpenAI-O1 has motivated a lot of people to think deeply about…thoughts 💭. Thinking before you speak is
        
     
    
            
    
        
    
        The Prompt Report Part 2: Plan and Solve, Tree of Thought, and Decomposition Prompting
            In the last blog, we went over prompting techniques 1-3 of The Prompt Report. This arXiv Dive, we were lucky
        
     
    
            
    
        
    
        The Prompt Report Part 1: A Systematic Survey of Prompting Techniques
            For this blog we are switching it up a bit. In past Arxiv Dives, we have gone deep into the
        
     
    
            
    
        
    
        ArXiv Dives: How ReFT works
            ArXiv Dives is a series of live meetups that take place on Fridays with the Oxen.ai community. We believe
        
     
    
            
    
        
    
        ArXiv Dives: Efficient DiT Fine-Tuning with PixART for Text to Image Generation
            Diffusion Transformers have been gaining a lot of steam since OpenAI's demo of Sora back in March. The
        
     
    
            
    
        
    
        ArXiv Dives: Evaluating LLMs for Code Completion with HumanEval
            Large Language Models have shown very good ability to generalize within a distribution, and frontier models have shown incredible flexibility
        
     
    
            
    
        
    
        How to Train Diffusion for Text from  Scratch
            This is part two of a series on Diffusion for Text with Score Entropy Discrete Diffusion (SEDD) models. Today we