Mariusz Kurman's Avatar

Mariusz Kurman

@mkurman

from bedside to byte_side, MD to AI, πŸ‡΅πŸ‡±

36
Followers
30
Following
22
Posts
27.11.2024
Joined
Posts Following

Latest posts by Mariusz Kurman @mkurman

Post image

The one thing I love is how Opus drafts its plan. Data flow is a masterpiece - and yes, I started to carefully read the details of what it's going to proceed, because A) it makes mistakes or misunderstands, B) it's highly educational.

11.01.2026 16:20 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Preview
GitHub - mkurman/synthlabs Contribute to mkurman/synthlabs development by creating an account on GitHub.

- Data Preview: Have data but unsure what's inside? Explore it directly!
- Verifier View: evaluate generated data, remove duplicates, assign ratings

and many more!

github.com/mkurman/synt...

11.01.2026 10:44 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

- Multi-turn Support: pass one DEEP run, let the model ask follow-up questions, and choose who should respond using SYNTH-like thinking
- Firebase/Firestore: download your data directly as a JSONL file or upload it to your Firestore (production mode)

11.01.2026 10:44 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

- Generator: create your own dataset from scratch
- Converter: use existing datasets (Hugging Face support) with reasoning traces to match our SYNTH style
- DEEP Mode: multiple agents working together in various configurations

11.01.2026 10:44 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image Post image Post image Post image

Would you like to build your own SYNTH-like datasets or contribute to the development of SYNTHLabs?

Now you can. SYNTHLabs is fully open and waiting for your contribution!

What's inside?

11.01.2026 10:44 πŸ‘ 1 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0

I'm back here after a huge break to check out how bsky works now πŸ˜€

11.01.2026 10:40 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

It has a minor bug that requires further fine-tuning (sometimes it starts with the <|python_tag|> instead of <Thought>.

04.01.2025 13:17 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Preview
mkurman/llama-3.2-MEDIT-3B-o1 Β· Hugging Face We’re on a journey to advance and democratize artificial intelligence through open source and open science.

Here is my experimental Llama 3.2 3B with o1-like thinking. It utilizes Thoughts when needed, so don't be surprised when it's not.

Enjoy!

Give some likes to make me feel better πŸ˜‚

huggingface.co/mkurman/llam...

04.01.2025 13:17 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Video thumbnail

storm.genie.stanford.edu - A great tool from Stanford for creating articles. For me, a stronger Gemini with Deep Thinking. Definitely worth trying!

01.01.2025 17:34 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Deepseek MTP is something you should definitely look at

28.12.2024 14:35 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Predicting the next token as a learning objective is insufficient for optimal LLM training.

28.12.2024 01:35 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

I will definitely give it a try!

04.12.2024 22:28 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

HDIC - How Do I Contribute?

A new technique we are working on seems to have a huge impact on language models' generative capabilities, allowing the layers to self-esteem their contribution to the final prediction.

04.12.2024 18:48 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

RIP JetBrains subscription ☠️ after six years, it became too heavy to use as a daily IDE. Iβ€˜m now on the VS Code team.

04.12.2024 10:04 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

What research tools would you recommend for searching and analyzing scientific papers?

03.12.2024 14:18 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Preview
meditsolutions/SmolLM2-MedIT-Upscale-2B Β· Hugging Face We’re on a journey to advance and democratize artificial intelligence through open source and open science.

PS. You can find the model mentioned here: huggingface.co/meditsolutio...

License: Apache 2.0​​​​​​​​​​​​​​​​ 4/4

02.12.2024 14:53 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Preview
meditsolutions/SmolLM2-MedIT-Upscale-2B Β· Hugging Face We’re on a journey to advance and democratize artificial intelligence through open source and open science.

the model achieved better results in IFEval and a higher overall average score in Open LLM Leaderboard

I consider this a big success πŸ˜‡, since surpassing the original in metrics is often very time-consuming, generates high costs, and doesn't always work out. 3/4

02.12.2024 14:53 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Preview
meditsolutions/SmolLM2-MedIT-Upscale-2B Β· Hugging Face We’re on a journey to advance and democratize artificial intelligence through open source and open science.

In total, not much really, since we don't have the original trained under the same conditions as our upscale. However...

1. We scaled up the model without losing its quality
2. We confirmed that the method we devised works
3. After extremely short fine-tuning, 2/4

02.12.2024 14:53 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Preview
meditsolutions/SmolLM2-MedIT-Upscale-2B Β· Hugging Face We’re on a journey to advance and democratize artificial intelligence through open source and open science.

We built a new small language model SmolLM2-MedIT-Upscale-2B, based on SmolLM2-1.7B-Instruct from Hugging Face. The premise was simple - increasing the vector in attention layers would positively impact the model's capabilities.

What did we prove? 1/4

02.12.2024 14:53 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
QUEBONAFIDE - FUTURAMA 3
QUEBONAFIDE - FUTURAMA 3 YouTube video by PÓŁNOC / POŁUDNIE

This video clip excellently demonstrates the exceptional uses of AI.

youtu.be/MMryYio0v6k?...

30.11.2024 23:28 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

It looks like we can scale up any model not only in-depth but also in width. Insane.
#llm

29.11.2024 11:26 πŸ‘ 3 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
SpeakLeash | Spichlerz Najnowsze wiadomoΕ›ci O nas Otwarty projekt, ktΓ³rego celem jest zbudowanie zestawu danych dla DuΕΌego Modelu JΔ™zykowego, o rozmiarze co najmniej 1TB, skΕ‚adajΔ…cego siΔ™ z rΓ³ΕΌnorodnych tekstΓ³w w jΔ™zyku pol...

Polish has a nice set of open data from the SpeakLeash initiative.

Link: speakleash.org

27.11.2024 21:29 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0