Lj Miranda's Avatar

Lj Miranda

@ljvmiranda

PhD student at the University of Cambridge https://ljvmiranda921.github.io

489
Followers
111
Following
20
Posts
13.08.2024
Joined
Posts Following

Latest posts by Lj Miranda @ljvmiranda

Finally, I want to thank the folks from HuggingFace for helping draft the official blog post (special shoutout to @clefourrier , @vanstriendaniel, @nathanhabib1011) and @Cohere_Labs for the research credits. :)

20.08.2025 20:40 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Evals are often the first step, we hope FilBench paves the way for language-specific adaptation especially for Philippine languages! I've written some of my thoughts here:

ljvmiranda921.github.io/projects/20...

20.08.2025 20:40 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Here's the link to the paper and leaderboard:

πŸ“œ Paper: arxiv.org/abs/2508.03523
πŸ“Š Leaderboard: ud-filipino-filbench-leaderboard.hf.space/

20.08.2025 20:40 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

This collaboration is exciting, it felt like assembling the Avengers of Filipino NLP. @acocodes and Conner are great collaborators, and I was happy to team-up with @jcblaisecruz and @josephimperial_, who are working on Filipino NLP for longer than I did!

20.08.2025 20:40 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Preview
πŸ‡΅πŸ‡­ FilBench - Can LLMs Understand and Generate Filipino?

πŸ‡΅πŸ‡­ One of my research interests is improving the state of Filipino NLP

Happy to share that we're taking a major step towards this by introducing FilBench, an LLM benchmark for Filipino!

Also accepted at EMNLP Main! πŸŽ‰

Learn more:
huggingface.co/blog/filbench

20.08.2025 20:40 πŸ‘ 4 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Preview
Field Report: ACL 2025 A collection of notes, projects, and essays.

ljvmiranda921.github.io/notebook/202...

02.08.2025 03:25 πŸ‘ 2 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image Post image

Ai2 is excited to be at #ACL2025 in Vienna, Austria this week. Come say hello, meet the team, and chat about the future of NLP. See you there! πŸ€πŸ“š

28.07.2025 17:00 πŸ‘ 9 πŸ” 3 πŸ’¬ 0 πŸ“Œ 0
Preview
Crowdsource, Crawl, or Generate? Creating SEA-VL, a Multicultural Vision-Language Dataset for Southeast Asia Samuel Cahyawijaya, Holy Lovenia, Joel Ruben Antony Moniz, Tack Hwa Wong, Mohammad Rifqi Farhansyah, Thant Thiri Maung, Frederikus Hudi, David Anugraha, Muhammad Ravi Shulthan Habibi, Muhammad Reza Qorib, Amit Agarwal, Joseph Marvin Imperial, Hitesh Laxmichand Patel, Vicky Feliren, Bahrul Ilmi Nasution, Manuel Antonio Rufino, Genta Indra Winata, Rian Adam Rajagede, Carlos Rafael Catalan, Mohamed Fazli Mohamed Imam, Priyaranjan Pattnayak, Salsabila Zahirah Pranida, Kevin Pratama, Yeshil Bangera, Adisai Na-Thalang, Patricia Nicole Monderin, Yueqi Song, Christian Simon, Lynnette Hui Xian Ng, Richardy Lobo Sapan, Taki Hasan Rafi, Bin Wang, Supryadi, Kanyakorn Veerakanjana, Piyalitt Ittichaiwong, Matthew Theodore Roque, Karissa Vincentio, Takdanai Kreangphet, Phakphum Artkaew, Kadek Hendrawan Palgunadi, Yanzhi Yu, Rochana Prih Hastuti, William Nixon, Mithil Bangera, Adrian Xuan Wei Lim, Aye Hninn Khine, Hanif Muhammad Zhafran, Teddy Ferdinan, Audra Aurora Izzani, Ayushman Singh, Evan Evan,

I was also part of a large-scale @seacrowd.bsky.social collaboration on building a vision-language dataset tailored for Southeast Asian Languages :) Also at ACL Main - aclanthology.org/2025.acl-lo...

July 29 Hall 4/5 10:30-12:00

#ACL2025 #ACL2025NLP

24.07.2025 12:56 πŸ‘ 2 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

3️⃣ The UD-NewsCrawl Treebank: Reflections and Challenges from a Large-scale Tagalog Syntactic Annotation Project (Main) -
aclanthology.org/2025.acl-lo...

July 29 Hall 4/5 10:30-12:00

Collab with folks from UP Diliman

#ACL2025 #ACL2025NLP

24.07.2025 12:56 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

2️⃣ M-RewardBench: Evaluating Reward Models in Multilingual Settings (Main) - aclanthology.org/2025.acl-lo...

July 28 Hall 4/5 11:00-12:30

Collab with folks from @cohereforai.bsky.social

#ACL2025 #ACL2025NLP

24.07.2025 12:56 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

1️⃣ Hybrid Preferences: Learning to Route Instances for Human vs. AI Feedback (Main) - aclanthology.org/2025.acl-lo...

7/29 Hall 4/5 10:30-12:00

My project here at @ai2.bsky.social!

#ACL2025NLP

24.07.2025 12:56 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

I'll be at @aclmeeting.bsky.social‬ in Vienna! I'm going to present the ff first/co-first author works:

24.07.2025 12:56 πŸ‘ 3 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Preview
β€˜Draw me a swordsman’: Can tool-calling LLMs draw pixel art? Just a fun weekend experiment on model-context protocol (MCP): I asked several tool-calling LLMs to draw a 4-frame spritesheet of a swordsman performing a sl...

fun learning stuff (+ phew i haven't blogged in a long time!): ljvmiranda921.github.io/notebook/202...

20.07.2025 04:36 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image Post image

We’re thrilled that SEA-VL has been accepted to the ACL 2025 (Main)!

Thank you to everyone who contributed to this project πŸ₯³

Paper: arxiv.org/abs/2503.07920
Project: seacrowd.github.io/seavl-launch/

#ACL2025NLP #SEACrowd #ForSEABySEA

16.05.2025 22:18 πŸ‘ 2 πŸ” 4 πŸ’¬ 0 πŸ“Œ 0
Image illustrating that ALM can enable Ensembling, Transfer to Bytes, and general Cross-Tokenizer Distillation.

Image illustrating that ALM can enable Ensembling, Transfer to Bytes, and general Cross-Tokenizer Distillation.

We created Approximate Likelihood Matching, a principled (and very effective) method for *cross-tokenizer distillation*!

With ALM, you can create ensembles of models from different families, convert existing subword-level models to byte-level and a bunch more🧡

02.04.2025 06:36 πŸ‘ 25 πŸ” 14 πŸ’¬ 1 πŸ“Œ 0
Post image

πŸ•΅πŸ»πŸ’¬ Introducing Feedback Forensics: a new tool to investigate pairwise preference data.

Feedback data is notoriously difficult to interpret and has many known issues – our app aims to help!

Try it at app.feedbackforensics.com

Three example use-cases πŸ‘‡πŸ§΅

17.03.2025 18:12 πŸ‘ 7 πŸ” 2 πŸ’¬ 1 πŸ“Œ 0
Preview
allenai/olmo-2-0325-32b-preference-mix Β· Datasets at Hugging Face We’re on a journey to advance and democratize artificial intelligence through open source and open science.

OLMo 2 0325 32B Preference Mixture: Solves AI alignment challenges through diverse preferences

- Combines 7 datasets
- Filters for instruction-following capability
- Balances on-policy and off-policy prompts
- Enabled successful DPO of OLMo-2-0325-32B model

huggingface.co/datasets/all...

13.03.2025 19:45 πŸ‘ 6 πŸ” 2 πŸ’¬ 1 πŸ“Œ 0
The logo for TΓΌlu 405B.

The logo for TΓΌlu 405B.

Here is Tülu 3 405B 🐫 our open-source post-training model that surpasses the performance of DeepSeek-V3! It demonstrates that our recipe, which includes RVLR scales to 405B - with performance on par with GPT-4o, & surpassing prior open-weight post-trained models of the same size including Llama 3.1.

30.01.2025 14:28 πŸ‘ 92 πŸ” 20 πŸ’¬ 2 πŸ“Œ 8
Post image

kicking off 2025 with our OLMo 2 tech report while payin homage to the sequelest of sequels 🫑

πŸš— 2 OLMo 2 Furious πŸ”₯ is everythin we learned since OLMo 1, with deep dives into:

πŸš– stable pretrain recipe
πŸš” lr anneal 🀝 data curricula 🀝 soups
🚘 tulu post-train recipe
🚜 compute infra setup

πŸ‘‡πŸ§΅

03.01.2025 16:02 πŸ‘ 69 πŸ” 17 πŸ’¬ 2 πŸ“Œ 1
Post image

BERT is BACK! I joined a collaboration with AnswerAI and LightOn to bring you the next iteration of BERT.

Introducing ModernBERT: 16x larger sequence length, better downstream performance (classification, retrieval), the fastest & most memory efficient encoder on the market.
🧡

19.12.2024 16:41 πŸ‘ 48 πŸ” 7 πŸ’¬ 1 πŸ“Œ 0
Preview
This is where the data to build AI comes from New findings show how the sources of data are concentrating power in the hands of the most powerful tech companies.

New research reveals a worrying trend: AI's data practices risk concentrating power overwhelmingly in the hands of dominant technology companies. With analysis from
@shaynelongpre.bsky.social @sarahooker.bsky.social @smw.bsky.social @giadapistilli.com www.technologyreview.com/2024/12/18/1...

18.12.2024 17:34 πŸ‘ 86 πŸ” 31 πŸ’¬ 2 πŸ“Œ 9
NeurIPS Tutorial Experimental Design and Analysis for AI ResearchersNeurIPS 2024

Stop by our #NeurIPS tutorial on Experimental Design & Analysis for AI Researchers! πŸ“Š

neurips.cc/virtual/2024/tutorial/99528

Are you an AI researcher interested in comparing models/methods? Then your conclusions rely on well-designed experiments. We'll cover best practices + case studies. πŸ‘‡

07.12.2024 18:15 πŸ‘ 86 πŸ” 13 πŸ’¬ 6 πŸ“Œ 1
Models The AI for Humanists project is developing resources to enable DH scholars to explore how large language models and AI technologies can be used in their research and teaching. Find an annotated biblio...

We just updated the AI for Humanists guide to model selection to include Llama 3.3, and a recommended best cost/capability tradeoff, llama 3.1 8B. What have you tried, and what would you suggest?

aiforhumanists.com/guides/models/

10.12.2024 19:04 πŸ‘ 54 πŸ” 19 πŸ’¬ 3 πŸ“Œ 0
NeurIPS Tutorial Opening the Language Model Pipeline: A Tutorial on Data Preparation, Model Training, and AdaptationNeurIPS 2024

the science of LMs should be fully open✨

today @akshitab.bsky.social @natolambert.bsky.social and I are giving our #neurips2024 tutorial on language model development.

everything from data, training, adaptation. published or not, no secrets 🫑

tues, 12/10, 9:30am PT β˜•οΈ

neurips.cc/virtual/2024...

10.12.2024 15:31 πŸ‘ 147 πŸ” 17 πŸ’¬ 5 πŸ“Œ 3

Come chat with me at #NeurIPS2024 and learn about how to use Paloma to evaluate perplexity over hundreds of domains! ✨We have stickers too✨

10.12.2024 03:54 πŸ‘ 21 πŸ” 4 πŸ’¬ 1 πŸ“Œ 0
Grassroots Science A global initiative focused on developing state-of-the-art multilingual language models through grassroots efforts.

⭐️ We're going to launch Grassroots Science, a year-long ambitious, massive-scale, fully open-source initiative aimed at developing multilingual LLMs aligned to diverse and inclusive human preferences in Feb 2025.

🌐 Check our website: grassroots.science.

#NLProc #GrassrootsScience

09.12.2024 05:02 πŸ‘ 7 πŸ” 5 πŸ’¬ 1 πŸ“Œ 3

Thank you @oxykodit.bsky.social !

04.12.2024 17:48 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Happy to share this and excited to bring this to the public! Nice collab with folks from the University of the Philippines (UP), @angelaquino_ph and Elsie Or, for this impactful work :) Hoping to have the official UD release next year as well.

04.12.2024 04:28 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Preview
UD-Filipino/UD_Tagalog-NewsCrawl Β· Datasets at Hugging Face We’re on a journey to advance and democratize artificial intelligence through open source and open science.

We're releasing the largest Universal Dependencies (UD) treebank for Tagalog, UD-NewsCrawl! This dataset has been a long time coming, but glad to see this through: 15k+ sentences versus the previous ~150 sents from older Tagalog treebanks.

πŸ€— : huggingface.co/datasets/UD-...
πŸ“ : Paper soon!

04.12.2024 04:28 πŸ‘ 14 πŸ” 2 πŸ’¬ 1 πŸ“Œ 1
Jonathan Berant (Tel Aviv University / Google) / Towards Robust Language Model Post-training
Jonathan Berant (Tel Aviv University / Google) / Towards Robust Language Model Post-training YouTube video by Yoav Artzi

I am seriously behind uploading Learning Machines videos, but I did want to get @jonathanberant.bsky.social's out sooner than later. It's not only a great talk, it also gives a remarkably broad overview and contextualization, so it's an excellent way to ramp up on post-training
youtu.be/2AthqCX3h8U

02.12.2024 03:45 πŸ‘ 53 πŸ” 12 πŸ’¬ 1 πŸ“Œ 0