Skip Navigation

InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)P
Posts
3
Comments
14
Joined
3 yr. ago

  • Ah. Sorry, good thing I attached the related link.

  • Everybody been rumoring about R2. So releasing this thing kinda unexpected

  • So something like

     
        
    Previously the text talk about [last summary]
    [The instruction prompt]...
    [Current chunk/paragraphs]
    
    
      
  • The RL is so good grok changed it's personality by changing small part of it's system prompt

  • Llama 3.3 was good, tho. For the multimodal, llama 4 also use llama3.2 approach where the image and text is made into single model instead using CLIP or siglip.

  • They got the whole Twitter database. It's kinda the same with Gemini. But somehow Meta isn't catching up, maybe their llama 4 architecture isn't that stable to train.

  • It changed after Grok 3

  • LocalLLaMA @sh.itjust.works

    When DeepSeek V4 and R2?

  • Lots of developer choose to write in CUDA as ROCm support back then is a mess.

  • No, you can run sd, flux based model inside the koboldcpp. You can try it out using the original koboldcpp in google colab. It loads gguf model. Related discussion on Reddit: https://www.reddit.com/r/StableDiffusion/comments/1gsdygl/koboldcpp_now_supports_generating_images_locally/

    Edit: Sorry, I kinda missed the point, maybe I'm sleepy when writing that comment. Yeah, I agree that LLM need big memory to run which is one of it's downside. I remember someone doing comparison that API with token based pricing is cheaper that to run it locally. But, running image generation locally is cheaper than API with step+megapixel pricing.

  • Skywork downfall

  • LocalLLaMA @sh.itjust.works

    MindLink-32B and MindLink-72B available on Huggingface