Data Science

  • 202 Posts
  • 748 Comments
Joined 2 years ago
cake
Cake day: June 17th, 2023

help-circle










  • From the article:

    DeepSeek-R1 release leaves open several questions about:

    • Data collection: How were the reasoning-specific datasets curated?
    • Model training: No training code was released by DeepSeek, so it is unknown which hyperparameters work best and how they differ across different model families and scales.
    • Scaling laws: What are the compute and data trade-offs in training reasoning models?

    These questions prompted us to launch the Open-R1 project, an initiative to systematically reconstruct DeepSeek-R1’s data and training pipeline, validate its claims, and push the boundaries of open reasoning models. By building Open-R1, we aim to provide transparency on how reinforcement learning can enhance reasoning, share reproducible insights with the open-source community, and create a foundation for future models to leverage these techniques.

    In this blog post we take a look at key ingredients behind DeepSeek-R1, which parts we plan to replicate, and how to contribute to the Open-R1 project







  • Embedded software development has dramatically advanced over the past decade. What does that mean for bare-metal programming?

    At a Glance

    • Bare-metal programming is an essential skill as it enables you to understand what your system is doing at the lowest levels.
    • Even if you spend your days working with abstraction layers, bare-metal programming will guide you should abstractions fail.
    • And bare-metal skills can provide a solid foundation for troubleshooting and debugging.