-
LLMs Can Get "Brain Rot"!
Paper • 2510.13928 • Published • 23 -
AmberYifan/llama3-8b-full-pretrain-mix-low-tweet-1m-en-gpt-sft
Text Generation • 8B • Updated • 1 -
AmberYifan/llama3-8b-full-pretrain-mix-mid-tweet-1m-en-gpt-sft
Text Generation • 8B • Updated • 2 -
AmberYifan/llama3-8b-full-pretrain-mix-high-tweet-1m-en-gpt-sft
Text Generation • 8B • Updated • 3
Shuo Xing
shuoxing
AI & ML interests
MLLMs, LLMs
Recent Activity
updated a collection 1 day ago
LLMs Can Get “Brain Rot”: A Pilot Study on Twitter/X updated a model 1 day ago
shuoxing/llama3-8b-full-pretrain-mix-low-tweet-1m-en updated a model 1 day ago
shuoxing/llama3-8b-full-pretrain-mix-low-tweet-1m-en-sftOrganizations
LLM4Math
-
BrokenMath: A Benchmark for Sycophancy in Theorem Proving with LLMs
Paper • 2510.04721 • Published -
FormalMATH: Benchmarking Formal Mathematical Reasoning of Large Language Models
Paper • 2505.02735 • Published • 33 -
PolyMath: Evaluating Mathematical Reasoning in Multilingual Contexts
Paper • 2504.18428 • Published -
MathConstruct: Challenging LLM Reasoning with Constructive Proofs
Paper • 2502.10197 • Published
LLMs Can Get “Brain Rot”: A Pilot Study on Twitter/X
-
LLMs Can Get "Brain Rot"!
Paper • 2510.13928 • Published • 23 -
AmberYifan/llama3-8b-full-pretrain-mix-low-tweet-1m-en-gpt-sft
Text Generation • 8B • Updated • 1 -
AmberYifan/llama3-8b-full-pretrain-mix-mid-tweet-1m-en-gpt-sft
Text Generation • 8B • Updated • 2 -
AmberYifan/llama3-8b-full-pretrain-mix-high-tweet-1m-en-gpt-sft
Text Generation • 8B • Updated • 3
LLM4Math
-
BrokenMath: A Benchmark for Sycophancy in Theorem Proving with LLMs
Paper • 2510.04721 • Published -
FormalMATH: Benchmarking Formal Mathematical Reasoning of Large Language Models
Paper • 2505.02735 • Published • 33 -
PolyMath: Evaluating Mathematical Reasoning in Multilingual Contexts
Paper • 2504.18428 • Published -
MathConstruct: Challenging LLM Reasoning with Constructive Proofs
Paper • 2502.10197 • Published
models 227
shuoxing/llama3-8b-full-pretrain-mix-low-tweet-1m-en
Text Generation • 266k • Updated • 15
shuoxing/llama3-8b-full-pretrain-mix-low-tweet-1m-en-sft
Text Generation • 266k • Updated • 16
shuoxing/llama3-8b-full-pretrain-mix-mid-tweet-1m-en
Text Generation • 266k • Updated • 11
shuoxing/llama3-8b-full-pretrain-mix-mid-tweet-1m-en-sft
Text Generation • 266k • Updated • 14
shuoxing/llama3-8b-full-pretrain-mix-high-tweet-1m-en
Text Generation • 266k • Updated • 11
shuoxing/llama3-8b-full-pretrain-mix-high-tweet-1m-en-sft
Text Generation • 266k • Updated • 11
shuoxing/llama3-8b-full-pretrain-control-tweet-1m-en-sft
Text Generation • 266k • Updated • 6
shuoxing/qwen2.5-0.5b-instruct-full-pretrain-mix-low-tweet-1m-en-sft
Text Generation • 0.5B • Updated • 13
shuoxing/qwen2.5-0.5b-instruct-full-pretrain-mix-mid-tweet-1m-en-sft
Text Generation • 0.5B • Updated • 11
shuoxing/qwen2.5-0.5b-instruct-full-pretrain-mix-high-tweet-1m-en-sft
Text Generation • 0.5B • Updated • 16
datasets 7
shuoxing/yt_ugc_public
Updated • 1.12k
shuoxing/AutoTrust
Updated • 10
shuoxing/KoNViD_1k_videos
Viewer • Updated • 1.2k • 178
shuoxing/Tweet_demo
Viewer • Updated • 100 • 13
shuoxing/MapBench_VQA
Viewer • Updated • 96 • 49 • 1
shuoxing/MapBench
Viewer • Updated • 97 • 11
shuoxing/tweet-scholar
Viewer • Updated • 95 • 14