Deconstructing the evaluation system through data analysis
Every year, numerous university rankings are published globally.
The Times Higher Education (THE) ranking is one of the most influential.
The official indicator weights are public. But does the data actually match the methodology?
Source Dataset: 2,603 records | 818 Universities | 72 Countries
Analysis Focus: 2015 dataset (201 top universities).
Preparation:
The distribution of top universities is not even.
It shows massive geographical concentration:
How do the indicators behave statistically?
How can actual empirical weights be found?
The final score \(S\) must be treated as a mathematical function of five variables.
By applying linear regression, it is possible to extract the gradient components:
\[\nabla S = \left( \frac{\partial S}{\partial Teach}, \frac{\partial S}{\partial Res}, \frac{\partial S}{\partial Cit}, \frac{\partial S}{\partial Int}, \frac{\partial S}{\partial Inc} \right)\]
Is the THE ranking a “black box”?
No. The empirical calculations show near-perfect agreement with the stated methodology.
| Indicator | Empirical Weight | Official Weight |
|---|---|---|
| Teaching | \(\approx\) 30% | 30% |
| Research | \(\approx\) 30% | 30% |
| Citations | \(\approx\) 30% | 30% |
| International | \(\approx\) 7.5% | 7.5% |
| Income | \(\approx\) 2.5% | 2.5% |
The relationship is strictly linear for the “Heavyweight Pillars”.
International and Income show much higher dispersion.
1. Transparency Confirmed The system is mathematically predictable. If you know the formula, you know the rank.
2. Strategic Focus To reach the top, a university must excel in the “Big Three” (Teaching, Research, Citations) — they control 90% of the outcome.
3. Geographical Bias The current 30/30/30 structure inherently favors the Anglo-Saxon university model, cementing the US and UK dominance.