Overview

  • Sectors Distribution/Logistics
  • Posted Jobs 0
  • Viewed 15

Company Description

Open-R1: a Fully Open Reproduction Of DeepSeek-R1

Hey there! This blog site post is an intro to the task, not a claim that we have actually R1 yet. We’re constructing in the open, so as soon as we have examination numbers, we’ll share them. You can follow our progress on Hugging Face and GitHub.

True, but it looks like there’s absolutely nothing to be examined as of today. I assume the ultimate objective is to train a new thinking model and then use the same assessment metrics as o1 and the DeepSeek-R1.

Well, there need to be at least some sanity check and validation to ensure the design was trained correctly.

Oh yes, if you are speaking about the evaluation variety of deepseek’s model it’s coming soon!

As mentioned in the blog post there is no design called Open-R1 to check at all … not yet anyhow. This is a blog site outlining that Hugging face will take the R1 Deepseek model, exercise how it was constructed as outlined in the paper and from what they released, and then reproduce that procedure.

in reality this is practically how science works … A creates a plan, discovery or development and it is checked by B, C and D to see if it is reproduceable. Thats been the cornerstone of research now for a few centuries.

This blog site is not saying they have actually already done so … Its a blog describing an intent to start training a design like R1 and calling it Open-R1.

Also DeepSeek-R1 was just released recently, and even in their paper they detailed the compute hours required. While those are low calculate hours for a SOTA model this does not mean you can train said model in a week. I ‘d personally enjoy to be able to train a transformer model in a week, however we may need to wait a while for that level of calculate technology.

So there are no benchmarks for a model that has not been built yet right? As described in the blog site, and again in reply to your concern.

However fear not, there is a GitHub Repo already and contributors (hell I may join myself), some prelim work done, and a master plan. A good beginning position.

n
@edbeeching
has evaluated the released designs already

( src: https://x.com/edwardbeeching/status/1884273209136275742)

R1 just trained on o1 outputs, so collectively …/ s. This is what the brand-new AI czars are stating

Hi! This article is an introduction to the job, not a claim that we’ve reproduced R1 yet. We will completely share the missing piece when we have them, you can anticipate the designs and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo

That’s nice and essential to understand this remarkable hype that lacks technical understanding and explanation. Science is about recreation, and if they claim to be open, let them fullfill the open part.

Please do publish the training expense.

We will!

Excalidraw Hi n
@bojan2501
thanks, we will indeed be working hard to make sure this training recipe can work for small language designs on consumer hardware because not everybody has a cluster of H100s in your home:-RRB- The tool we used for the images was Excalidraw! https://excalidraw.com

looking forward to it! WTF are your talking about?

need to be a joke

It’s truly cool to see how the whole open source community comes together!

Ops …

5.5 M is number reporter in the deepseekv3 tech report (just the training, not the experiment afaik), for R1 difficult to estimate tbh however much less than 5.5 M imo

Historically, they have actually never ever launched code or datasets of their LLM training, so I wouldn’t anticipate this time to be various. If they would launch it that would be amazing naturally!

Yes of course!

So essentially you’re asking to change existing censorship with another flavour of censorship?

The code for the designs are inside the design repositories, e.g. for V3: https://huggingface.co/deepseek-ai/DeepSeek-V3/blob/main/modeling_deepseek.py

Hello Team, I’m Ray Bernard, the author and creator of EQUATOR. My research study group will be working on a paper concentrated on replicating specific components of DeepSeek R1. Our aim is to recreate the cold start and provide your group with a dataset that consists of COT and other strategies to support these efforts. We like to contribute our work to assist. Please let me understand if you find this useful. Best, Ray Bernard https://www.facebook.com/groups/1186310571520299/

Where is the evaluation numbers? without it you can’t call it reproduction.

8 replies

True, however it appears like there’s nothing to be examined as of today. I assume the supreme goal is to train a brand-new thinking design and after that utilize the same evaluation metrics as o1 and the DeepSeek-R1.

That’s rather intriguing, I was asking myself why the questions the author exposed here are not being asked by others? I believe the work they have actually done is remarkable but at the exact same time I wonder why they would not put these missing out on pieces on if they are supposed to be totally open.
Why even without reproduction and comprehension of the development they could impact so much the market in this method?

4 replies

Hi! This article is an intro to the job, not a claim that we’ve reproduced R1 yet. We will completely share the missing out on piece when we have them, you can expect the designs and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo

Interesting read, and it is excellent that we see more effort into this direction: more optimization and less brute force.
Also wonder what tool did the author usage for creating step diagram.

2 replies

Excalidraw I’m so happy that effort like this already exist, I’m gon na attempt to contribute:-RRB- 1 reply

looking forward to it! So racist articel

2 replies

WTF are your discussing?

Awesome to have this open recreation started!

For Step # 1 check out https://github.com/open-thoughts/open-thoughts!

https://x.com/ryanmart3n/status/1884284101265612856

Let’s do this thing!

1 reply

It’s actually cool to see how the entire open source community comes together!

Does anyone understand the real training cost of r1? I can’t find it in the paper or the statement post. Is the 6M expense reported by media just the number drawn from v3’s training expense?

2 replies

Ops …

Has anybody asked the DeepSeek team to release their training data and code, or at least share them privately with an independent replication project like this? Have they declined such a demand?

A loyal duplication depends upon utilizing the exact same dataset and hyperparameters. Otherwise, any significant disparities with the published criteria would be hard to pin down-whether due to training information differences or the replication approach itself.

1 reply

Historically, they have never ever launched code or datasets of their LLM training, so I wouldn’t expect this time to be different. If they would launch it that would be amazing naturally!

In the meantime we have to make finest guess quotes and see if we can get there ourselves.

You provide excellent duplication process of Deepseek thinking training. I will try something similar to it.

This is truly great information, can we fine tune with particular use case when code is released?

1 reply

Yes naturally!

Please think about removing biased, tainted or unaligned training information and make an effort to get rid of copyrighted works from the crawl from consumption. This will make the model more usable. If you recycled anthropic curation checks, this might likewise assist, remove obviouslybiased information will likely add a lot of worth. We don’t desire another tainted, unaligned open source design, right? And no business would ever utilize deepseek or a design that recycles it, right?
We value your work for the benefit of mankind, we hope.
Miike C from NJ

1 reply

So generally you’re asking to replace existing censorship with another flavour of censorship?

Can’t wait! Hopefully the model will be uncensored but whatever you can do is alright! Love seeing open source structure itself up. I’m not wise enough to actually assist but I can contribute moral assistance lol

Hello guys, I am even simply searching for code for DeepSeek-V2, in order to completely comprehend multi-head latent attention. You do not appear to have code in Hugging Face even for that. Or am I missing out on something? Don’t see anything in src/transformers/models. MLA is not appropriately described in their paper, so it would be very important to have code for this.