π Data Science Riddle
Model Accuracy improves after dropping half the features. Why?
Model Accuracy improves after dropping half the features. Why?
Anonymous Quiz
12%
Model became smaller
71%
Overfitting reduced
11%
Data size shrank
6%
Training faster
β€3
Understanding the Forecast Statistics and Four Moments (4P).pdf
181.8 KB
Statistical Moments (M1, M2) for Data Analysis
Here are 5 curated PDFs diving into the mean (M1), variance (M2), and their applications in crafting research questions and sourcing data.
A channel member requested resources on this topic and we delivered.
If you have a topic you want resources on let us know, and weβll make it happen!
@datascience_bds
Here are 5 curated PDFs diving into the mean (M1), variance (M2), and their applications in crafting research questions and sourcing data.
A channel member requested resources on this topic and we delivered.
If you have a topic you want resources on let us know, and weβll make it happen!
@datascience_bds
β€8
π Data Science Riddle
Why do we use Batch Normalization?
Why do we use Batch Normalization?
Anonymous Quiz
31%
Speeds up training
41%
Prevents overfitting
7%
Adds non-linearity
21%
Reduces dataset size
β€3
π Data Science Riddle
Your object detection model misses small objects. Easiest fix?
Your object detection model misses small objects. Easiest fix?
Anonymous Quiz
22%
Use larger input images
30%
Add more classes
33%
Reduce learning rate
15%
Train longer
π€ AI that creates AI: ASI-ARCH finds 106 new SOTA architectures
ASI-ARCH β experimental ASI that autonomously researches and designs neural nets. It hypothesizes, codes, trains & tests models.
π‘ Scale:
1,773 experiments β 20,000+ GPU-hours.
Stage 1 (20M params, 1B tokens): 1,350 candidates beat DeltaNet.
Stage 2 (340M params): 400 models β 106 SOTA winners.
Top 5 trained on 15B tokens vs Mamba2 & Gated DeltaNet.
π Results:
PathGateFusionNet: 48.51 avg (Mamba2: 47.84, Gated DeltaNet: 47.32).
BoolQ: 60.58 vs 60.12 (Gated DeltaNet).
Consistent gains across tasks.
π Insights:
Prefers proven tools (gating, convs), refines them iteratively.
Ideas come from: 51.7% literature, 38.2% self-analysis, 10.1% originality.
SOTA share: self-analysis β to 44.8%, literature β to 48.6%.
@datascience_bds
ASI-ARCH β experimental ASI that autonomously researches and designs neural nets. It hypothesizes, codes, trains & tests models.
π‘ Scale:
1,773 experiments β 20,000+ GPU-hours.
Stage 1 (20M params, 1B tokens): 1,350 candidates beat DeltaNet.
Stage 2 (340M params): 400 models β 106 SOTA winners.
Top 5 trained on 15B tokens vs Mamba2 & Gated DeltaNet.
π Results:
PathGateFusionNet: 48.51 avg (Mamba2: 47.84, Gated DeltaNet: 47.32).
BoolQ: 60.58 vs 60.12 (Gated DeltaNet).
Consistent gains across tasks.
π Insights:
Prefers proven tools (gating, convs), refines them iteratively.
Ideas come from: 51.7% literature, 38.2% self-analysis, 10.1% originality.
SOTA share: self-analysis β to 44.8%, literature β to 48.6%.
@datascience_bds