oscarmoxon's comments

oscarmoxon | 2 days ago | on: Open Weights isn't Open Training

Some of this exists already in pockets (Common Crawl, The Pile, RedPajama are all volunteer/open efforts). I suppose there's no equivalent of the "edit this page and see the impact" like with have with Wikipedia. Contributing to an open dataset has no feedback loop if the training infrastructure that would consume it is closed... seems like a feedback problem.

oscarmoxon | 2 days ago | on: Open Weights isn't Open Training

Agree that this makes it unlikely we see frontier training data OS'd but this is a separate problem from software and infrastructure transparency, which has none of those constraints. Training stack, the parallelism decisions, documented failure modes are engineering knowledge and there's no principled reason it doesn't ship.

oscarmoxon | 2 days ago | on: Open Weights isn't Open Training

Agree, this feels like a distinction that needs formalising...

Passive transparency: training data, technical report that tells you what the model learned and why it behaves the way it does. Useful for auditing, AI safety, interoperability.

Active transparency: being able to actually reproduce and augment the model. For that you need the training stack, curriculum, loss weighting decisions, hyperparameter search logs, synthetic data pipeline, RLHF/RLAIF methodology, reward model architecture, what behaviours were targeted and how success was measured, unpublished evals, known failure modes. The list goes on!

oscarmoxon | 2 days ago | on: Open Weights isn't Open Training

Compute costs are falling fast, training is getting cheaper. GPT-2 costs pocket change to train, and now it costs pocket train to tune >1T parameter models. If it was transparent what costs went into the weights, they could be commodified and stripped of bloat. Instead the hidden cost is building the infrastructure that was never tested at scale by anyone other than the original developers who shipped no documentation of where it fails. Unlike compute, this hidden cost doesn't commodify on its own.

oscarmoxon | 3 days ago | on: Open Weights isn't Open Training

The framing here is undersold in the broader discourse: "open weights" is a ruse for reproducibility. What you have is closer to a compiled binary than source code. You can run it, you can diff it against other binaries, but you cannot, in any meaningful sense, reproduce or extend it from first principles.

This matters because OSS truly depends on the reproducibility claim. "Open weights" borrows the legitimacy of open source (the assumption that scrutiny is possible, that no single actor has a moat, that iteration is democratised). Truly democratised iteration would crack open the training stack and let you generate intelligence from scratch.

Huge kudos to Addie and the team for this :)

oscarmoxon | 9 months ago | on: Mullvad Leta

They're also littering the London tube system with ads - there's definitely been a lottery win or a series A.

oscarmoxon | 1 year ago | on: Overlay hundred-year-old British maps on modern satellite imagery

The National Library of Scotland has an impressive web tool that allows you to overlay detailed historical Ordnance Survey (OS) maps (dating back to 1841 for this instance) onto modern satellite imagery. The interface lets you adjust transparency and blend between past and present landscapes.

Here is the tool overlaid on Bath.

page 1