Gropromotions

Overview

  • Founded Date September 3, 1972
  • Sectors Non-Medical Non-Clinical
  • Posted Jobs 0
  • Viewed 34
Bottom Promo

Company Description

Open-R1: a Completely Open Reproduction Of DeepSeek-R1

Hey there! This article is an intro to the task, not a claim that we have actually replicated R1 yet. We’re constructing in the open, so as quickly as we have evaluation numbers, we’ll share them. You can follow our development on Hugging Face and GitHub.

True, but it appears like there’s nothing to be examined as of today. I presume the supreme goal is to train a brand-new reasoning design and then use the very same examination metrics as o1 and the DeepSeek-R1.

Well, there must be at least some peace of mind check and validation to make sure the design was trained correctly.

Oh yes, if you are discussing the evaluation variety of deepseek’s design it’s coming soon!

As pointed out in the article there is no design called Open-R1 to check at all … not yet anyway. This is a blog site laying out that Hugging face will take the R1 Deepseek model, exercise how it was constructed as detailed in the paper and from what they released, and after that replicate that procedure.

in reality this is quite much how science works … A develops a plan, discovery or development and it is tested by B, C and D to see if it is reproduceable. Thats been the cornerstone of research now for a couple of centuries.

This blog site is not stating they have actually currently done so … Its a an intent to begin training a design like R1 and calling it Open-R1.

Also DeepSeek-R1 was only launched last week, and even in their paper they outlined the calculate hours needed. While those are low compute hours for a SOTA model this does not indicate you can train stated model in a week. I ‘d personally enjoy to be able to train a transformer model in a week, however we may require to wait a while for that level of compute innovation.

So there are no benchmarks for a design that has not been built yet right? As laid out in the blog site, and again in reply to your concern.

However fear not, there is a GitHub Repo currently and factors (hell I might join myself), some prelim work done, and a master plan. An excellent starting position.

n
@edbeeching
has examined the launched designs already

( src: https://x.com/edwardbeeching/status/1884273209136275742)

R1 just trained on o1 outputs, so jointly …/ s. This is what the brand-new AI czars are saying

Hi! This post is an introduction to the project, not a claim that we have actually recreated R1 yet. We will absolutely share the missing piece when we have them, you can anticipate the models and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo

That’s great and essential to comprehend this significant hype that does not have technical comprehension and explanation. Science is about recreation, and if they declare to be open, let them fullfill the open part.

Please do publish the training cost.

We will!

Excalidraw Hi n
@bojan2501
thanks, we will undoubtedly be working hard to make sure this training recipe can work for little language designs on customer hardware since not everyone has a cluster of H100s in your home:-RRB- The tool we used for the images was Excalidraw! https://excalidraw.com

looking forward to it! WTF are your speaking about?

must be a joke

It’s truly cool to see how the whole open source community comes together!

Ops …

5.5 M is number reporter in the deepseekv3 tech report (simply the training, not the experiment afaik), for R1 hard to estimate tbh however much less than 5.5 M imo

Historically, they have never ever released code or datasets of their LLM training, so I would not anticipate this time to be different. If they would launch it that would be remarkable naturally!

Yes naturally!

So essentially you’re asking to replace existing censorship with another flavour of censorship?

The code for the designs are inside the design repositories, e.g. for V3: https://huggingface.co/deepseek-ai/DeepSeek-V3/blob/main/modeling_deepseek.py

Hello Team, I’m Ray Bernard, the author and creator of EQUATOR. My research group will be working on a paper focused on reproducing specific parts of DeepSeek R1. Our aim is to recreate the cold start and supply your group with a dataset that consists of COT and other techniques to support these efforts. We like to contribute our work to assist. Please let me understand if you discover this useful. Best, Ray Bernard https://www.facebook.com/groups/1186310571520299/

Where is the assessment numbers? without it you can’t call it reproduction.

8 replies

True, but it looks like there’s absolutely nothing to be assessed as of today. I presume the supreme goal is to train a brand-new reasoning design and after that use the exact same evaluation metrics as o1 and the DeepSeek-R1.

That’s quite intriguing, I was asking myself why the concerns the author exposed here are not being asked by others? I believe the work they have done is remarkable but at the very same time I wonder why they would not put these missing pieces on if they are supposed to be fully open.
Why even without recreation and understanding of the innovation they could impact a lot the market in this method?

4 replies

Hi! This post is an intro to the task, not a claim that we’ve recreated R1 yet. We will totally share the missing out on piece when we have them, you can anticipate the designs and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo

Interesting read, and it is good that we see more effort into this direction: more optimization and less strength.
Also wonder what tool did the author use for producing step diagram.

2 replies

Excalidraw I’m so glad that effort like this currently exist, I’m gon na attempt to contribute:-RRB- 1 reply

looking forward to it! So racist articel

2 replies

WTF are your discussing?

Awesome to have this open recreation began!

For Step # 1 check out https://github.com/open-thoughts/open-thoughts!

https://x.com/ryanmart3n/status/1884284101265612856

Let’s do this thing!

1 reply

It’s really cool to see how the entire open source community comes together!

Does anyone understand the actual training cost of r1? I can’t discover it in the paper or the statement post. Is the 6M expense reported by media just the number taken from v3‘s training expense?

2 replies

Ops …

Has anyone asked the DeepSeek team to publish their training data and code, or at least share them independently with an independent duplication job like this? Have they rejected such a request?

A devoted duplication depends on using the same dataset and hyperparameters. Otherwise, any major disparities with the published criteria would be tough to pin down-whether due to training information distinctions or the duplication technique itself.

1 reply

Historically, they have actually never ever released code or datasets of their LLM training, so I wouldn’t expect this time to be various. If they would launch it that would be remarkable obviously!

In the meantime we need to make best guess estimates and see if we can get there ourselves.

You supply great replication process of Deepseek reasoning training. I will try something similar to it.

This is truly excellent information, can we tweak with specific use case when code is launched?

1 reply

Yes obviously!

Please think about eliminating biased, tainted or unaligned training data and make an effort to remove copyrighted works from the crawl from consumption. This will make the design more functional. If you reused anthropic curation checks, this may likewise help, get rid of obviouslybiased data will likely add a great deal of value. We don’t want another polluted, unaligned open source model, right? And no business would ever utilize deepseek or a design that reuses it, right?
We appreciate your work for the benefit of humankind, we hope.
Miike C from NJ

1 reply

So basically you’re asking to change existing censorship with another flavour of censorship?

Can’t wait! Hopefully the model will be uncensored however whatever you can do is alright! Love seeing open source structure itself up. I’m not wise adequate to in fact help however I can contribute support lol

Hello guys, I am even simply trying to discover code for DeepSeek-V2, in order to completely understand multi-head latent attention. You do not seem to have code in Hugging Face even for that. Or am I missing something? Don’t see anything in src/transformers/models. MLA is not correctly described in their paper, so it would be very important to have code for this.

Bottom Promo
Bottom Promo
Top Promo