Sponsor of the Day:
Jerkmate
https://deepignorance.ai/
Deep Ignorance: Filtering Pretraining Data Builds Tamper-Resistant Safeguards into Open-Weight LLMs
Filtering pretraining can prevent unsafe knowledge, doesn’t sacrifice general performance, and results in models that are resistant to tampering.
open weight llmspretraining datatamper resistantdeepignorance
https://www.semanticscholar.org/search?q=Tag%26Tab%3A+Pretraining+Data+Detection+in+Large+Language+Models+Using+Keyword-Based+Membership+Inference+Attack.
Tag&Tab: Pretraining Data Detection in Large Language Models Using Keyword-Based Membership...
An academic search engine that utilizes artificial intelligence methods to provide highly relevant results and novel tools to filter them with ease.
large language modelspretraining databased membershiptagtab
https://arxiv.org/abs/2508.06601
[2508.06601] Deep Ignorance: Filtering Pretraining Data Builds Tamper-Resistant Safeguards into...
Abstract page for arXiv paper 2508.06601: Deep Ignorance: Filtering Pretraining Data Builds Tamper-Resistant Safeguards into Open-Weight LLMs
pretraining datatamper resistant2508deepignorance
https://research.google/pubs/analyzing-similarity-metrics-for-data-selection-for-language-model-pretraining/
Analyzing Similarity Metrics for Data Selection for Language Model Pretraining
language modelanalyzingsimilaritymetricsdata