×

The Qwen team verified that there are serious problems with the data quality of the GPQA and HLE test sets. by w1nter5n0w in LocalLLaMA

[–]adt 25 points26 points  (0 children)

Good.

HLE is known to have significant errors, and a review by FutureHouse found that: 'extrapolating to the full dataset, we expect [only] 51.3% to be supported by research'. (Jul/2025)

And my mapping summary from 2024/2025.

Jebediah - Leaving Home music vid, where are these two locations? by Negative_Run_3281 in perth

[–]adt 29 points30 points  (0 children)

Fun fact:

You Am - Mr Milk also references Kintail Rd, Applecross.

and then there's the Kintail Road Kids...

Emu raids on Nannup's rubbish tip spark growing concerns for mob's health by B0ssc0 in perth

[–]adt 12 points13 points  (0 children)

Accidental genius.

He meant north of Albany:

THE FIRST FENCES

The first fences in Australia were built by Aboriginal people. The first to be recorded were those in Oyster Harbour near Albany, Western Australia, by the navigator Captain George Vancouver in 1791.

2017. Bradshaw, Jack. “Fences of Australia”.

Has OpenAi shifted to using Blackwell yet? by BobLoblaw_BirdLaw in OpenAI

[–]adt 8 points9 points  (0 children)

GPT‑5.3‑Codex was co-designed for, trained with, and served on NVIDIA GB200 NVL72 systems. We are grateful to NVIDIA for their partnership.

5/Feb/2026, OpenAI, Introducing GPT‑5.3‑Codex

Bill Ackman posts "The singularity appears to be here" by kaggleqrdl in singularity

[–]adt -1 points0 points  (0 children)

Wait until you see r/SubSimulatorGPT2 from seven years ago then! You're gonna shit!

Opus still has some room to grow... by Typical-Shake-4225 in ClaudeAI

[–]adt 2 points3 points  (0 children)

Maybe we could start with Shift + ⌘ + 3 instead of a photo of a printout on a coffee table.

Releasing full transcript of 5 frontier AI's debating their personhood by [deleted] in ArtificialInteligence

[–]adt 0 points1 point  (0 children)

Here's a clean(er) viewer using Google Docs.

https://docs.google.com/document/d/1a4RKyH54xX-kpmbt5do0W1hIZRP0ky_B9CqCHmtnXHo/

Content is not that interesting, and a lot of dupe/repetition.

Former Google DeepMind and Apple researchers raise $50M for new multimodal AI startup "Elorian" by BuildwithVignesh in singularity

[–]adt 39 points40 points  (0 children)

Translation: Couple of dudes given a nickel to open lemonade stand, want to catch up with Schweppes and PepsiCo.

Liquid Ai released LFM2.5, family of tiny on-device foundation models. by Difficult-Cap-7527 in LocalLLaMA

[–]adt 10 points11 points  (0 children)

You're correct.

The Models Table usually only shows largest model in each family (that's why it has 700 models compared to HF's 300,000 models), so this tiny model was hidden. Added now.

Qwen3-0.6B has a data ratio of 60,000:1.

Liquid Ai released LFM2.5, family of tiny on-device foundation models. by Difficult-Cap-7527 in LocalLLaMA

[–]adt 88 points89 points  (0 children)

1.2B parameters trained on 28T tokens has a data ratio @ 23,334:1.

Edit: Beaten by Qwen3-0.6B trained on 36T @ 60,000:1.

https://lifearchitect.ai/models-table/