|
|
|
|
@ -1,47 +1,56 @@
|
|
|
|
|
{
|
|
|
|
|
"rejected_claims": [
|
|
|
|
|
{
|
|
|
|
|
"filename": "rlcf-architecture-separates-ai-generation-from-human-evaluation-with-bridging-selection.md",
|
|
|
|
|
"filename": "rlcf-architecture-separates-ai-generation-from-human-evaluation-with-bridging-algorithm-selection.md",
|
|
|
|
|
"issues": [
|
|
|
|
|
"missing_attribution_extractor"
|
|
|
|
|
]
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
"filename": "bridging-based-consensus-mechanisms-risk-homogenization-toward-optimally-inoffensive-outputs.md",
|
|
|
|
|
"filename": "bridging-based-consensus-mechanisms-risk-homogenization-toward-optimally-inoffensive-content.md",
|
|
|
|
|
"issues": [
|
|
|
|
|
"no_frontmatter"
|
|
|
|
|
]
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
"filename": "human-rating-authority-in-ai-systems-preserves-alignment-by-keeping-value-judgment-in-human-hands.md",
|
|
|
|
|
"issues": [
|
|
|
|
|
"missing_attribution_extractor"
|
|
|
|
|
]
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
"filename": "human-rating-authority-as-alignment-mechanism-preserves-judgment-sovereignty-while-scaling-content-generation.md",
|
|
|
|
|
"filename": "stylistic-novelty-rewards-in-rlcf-balance-optimization-pressure-with-diversity-preservation.md",
|
|
|
|
|
"issues": [
|
|
|
|
|
"missing_attribution_extractor"
|
|
|
|
|
]
|
|
|
|
|
}
|
|
|
|
|
],
|
|
|
|
|
"validation_stats": {
|
|
|
|
|
"total": 3,
|
|
|
|
|
"total": 4,
|
|
|
|
|
"kept": 0,
|
|
|
|
|
"fixed": 12,
|
|
|
|
|
"rejected": 3,
|
|
|
|
|
"fixed": 14,
|
|
|
|
|
"rejected": 4,
|
|
|
|
|
"fixes_applied": [
|
|
|
|
|
"rlcf-architecture-separates-ai-generation-from-human-evaluation-with-bridging-selection.md:set_created:2026-03-15",
|
|
|
|
|
"rlcf-architecture-separates-ai-generation-from-human-evaluation-with-bridging-selection.md:stripped_wiki_link:democratic-alignment-assemblies-produce-constitutions-as-eff",
|
|
|
|
|
"rlcf-architecture-separates-ai-generation-from-human-evaluation-with-bridging-selection.md:stripped_wiki_link:community-centred-norm-elicitation-surfaces-alignment-target",
|
|
|
|
|
"rlcf-architecture-separates-ai-generation-from-human-evaluation-with-bridging-selection.md:stripped_wiki_link:rlhf-is-implicit-social-choice-without-normative-scrutiny.md",
|
|
|
|
|
"bridging-based-consensus-mechanisms-risk-homogenization-toward-optimally-inoffensive-outputs.md:set_created:2026-03-15",
|
|
|
|
|
"bridging-based-consensus-mechanisms-risk-homogenization-toward-optimally-inoffensive-outputs.md:stripped_wiki_link:universal-alignment-is-mathematically-impossible-because-Arr",
|
|
|
|
|
"bridging-based-consensus-mechanisms-risk-homogenization-toward-optimally-inoffensive-outputs.md:stripped_wiki_link:pluralistic-alignment-must-accommodate-irreducibly-diverse-v",
|
|
|
|
|
"bridging-based-consensus-mechanisms-risk-homogenization-toward-optimally-inoffensive-outputs.md:stripped_wiki_link:some-disagreements-are-permanently-irreducible-because-they-",
|
|
|
|
|
"human-rating-authority-as-alignment-mechanism-preserves-judgment-sovereignty-while-scaling-content-generation.md:set_created:2026-03-15",
|
|
|
|
|
"human-rating-authority-as-alignment-mechanism-preserves-judgment-sovereignty-while-scaling-content-generation.md:stripped_wiki_link:human-in-the-loop-at-the-architectural-level-means-humans-se",
|
|
|
|
|
"human-rating-authority-as-alignment-mechanism-preserves-judgment-sovereignty-while-scaling-content-generation.md:stripped_wiki_link:coding-agents-cannot-take-accountability-for-mistakes-which-",
|
|
|
|
|
"human-rating-authority-as-alignment-mechanism-preserves-judgment-sovereignty-while-scaling-content-generation.md:stripped_wiki_link:economic-forces-push-humans-out-of-every-cognitive-loop-wher"
|
|
|
|
|
"rlcf-architecture-separates-ai-generation-from-human-evaluation-with-bridging-algorithm-selection.md:set_created:2026-03-15",
|
|
|
|
|
"rlcf-architecture-separates-ai-generation-from-human-evaluation-with-bridging-algorithm-selection.md:stripped_wiki_link:democratic-alignment-assemblies-produce-constitutions-as-eff",
|
|
|
|
|
"rlcf-architecture-separates-ai-generation-from-human-evaluation-with-bridging-algorithm-selection.md:stripped_wiki_link:community-centred-norm-elicitation-surfaces-alignment-target",
|
|
|
|
|
"rlcf-architecture-separates-ai-generation-from-human-evaluation-with-bridging-algorithm-selection.md:stripped_wiki_link:rlhf-is-implicit-social-choice-without-normative-scrutiny.md",
|
|
|
|
|
"bridging-based-consensus-mechanisms-risk-homogenization-toward-optimally-inoffensive-content.md:set_created:2026-03-15",
|
|
|
|
|
"bridging-based-consensus-mechanisms-risk-homogenization-toward-optimally-inoffensive-content.md:stripped_wiki_link:universal-alignment-is-mathematically-impossible-because-Arr",
|
|
|
|
|
"bridging-based-consensus-mechanisms-risk-homogenization-toward-optimally-inoffensive-content.md:stripped_wiki_link:pluralistic-alignment-must-accommodate-irreducibly-diverse-v",
|
|
|
|
|
"bridging-based-consensus-mechanisms-risk-homogenization-toward-optimally-inoffensive-content.md:stripped_wiki_link:some-disagreements-are-permanently-irreducible-because-they-",
|
|
|
|
|
"human-rating-authority-in-ai-systems-preserves-alignment-by-keeping-value-judgment-in-human-hands.md:set_created:2026-03-15",
|
|
|
|
|
"human-rating-authority-in-ai-systems-preserves-alignment-by-keeping-value-judgment-in-human-hands.md:stripped_wiki_link:coding-agents-cannot-take-accountability-for-mistakes-which-",
|
|
|
|
|
"human-rating-authority-in-ai-systems-preserves-alignment-by-keeping-value-judgment-in-human-hands.md:stripped_wiki_link:human-in-the-loop-at-the-architectural-level-means-humans-se",
|
|
|
|
|
"stylistic-novelty-rewards-in-rlcf-balance-optimization-pressure-with-diversity-preservation.md:set_created:2026-03-15",
|
|
|
|
|
"stylistic-novelty-rewards-in-rlcf-balance-optimization-pressure-with-diversity-preservation.md:stripped_wiki_link:pluralistic-ai-alignment-through-multiple-systems-preserves-",
|
|
|
|
|
"stylistic-novelty-rewards-in-rlcf-balance-optimization-pressure-with-diversity-preservation.md:stripped_wiki_link:high-AI-exposure-increases-collective-idea-diversity-without"
|
|
|
|
|
],
|
|
|
|
|
"rejections": [
|
|
|
|
|
"rlcf-architecture-separates-ai-generation-from-human-evaluation-with-bridging-selection.md:missing_attribution_extractor",
|
|
|
|
|
"bridging-based-consensus-mechanisms-risk-homogenization-toward-optimally-inoffensive-outputs.md:missing_attribution_extractor",
|
|
|
|
|
"human-rating-authority-as-alignment-mechanism-preserves-judgment-sovereignty-while-scaling-content-generation.md:missing_attribution_extractor"
|
|
|
|
|
"rlcf-architecture-separates-ai-generation-from-human-evaluation-with-bridging-algorithm-selection.md:missing_attribution_extractor",
|
|
|
|
|
"bridging-based-consensus-mechanisms-risk-homogenization-toward-optimally-inoffensive-content.md:no_frontmatter",
|
|
|
|
|
"human-rating-authority-in-ai-systems-preserves-alignment-by-keeping-value-judgment-in-human-hands.md:missing_attribution_extractor",
|
|
|
|
|
"stylistic-novelty-rewards-in-rlcf-balance-optimization-pressure-with-diversity-preservation.md:missing_attribution_extractor"
|
|
|
|
|
]
|
|
|
|
|
},
|
|
|
|
|
"model": "anthropic/claude-sonnet-4.5",
|
|
|
|
|
|