作者:158文章网日期:
返回目录:作文写作
Dear Mr.Fallows:
I appreciate that,as the new editor of U.S.News & World Report,you have much to do at this moment.However,it is precisely because you are the new editor that I write to you,personally.
I emphasize you,because of your demonstrated willingness to examine journalism in the same way that journalism examines all other facets of society.And I say personally because my letter is for your consideration,and not a letter to the editor for publication.
My timing also is related to the recent appearance of the annual U.S.News "America's Best Colleges" rankings.As the president of a university that is among the top-ranked universities,I hope I have the standing to persuade you that much about these rankings - particularly their specious formulas and spurious precision - is utterly misleading.I wish I could forego this letter since,after all,the rankings are only another newspaper story.Alas,alumni,foreign newspapers,and many others do not bring a sense of perspective to the matter.
I am extremely skeptical that the quality of a university - any more than the quality of a magazine - can be measured statistically.However,even if it can,the producers of the U.S.News rankings remain far from discovering the method.Let me offer as prima facie evidence two great public universities:the University of Michigan-Ann Arbor and the University of California-Berkeley.These clearly are among the very best universities in America - one could make a strong argument for either in the top half-dozen.Yet,in the last three years,the U.S.News formula has assigned them ranks that lead many readers to infer that they are second rate:Michigan 21-24-24,and Berkeley 23-26-27.
Such movement itself - while perhaps good for generating attention and sales - corrodes the credibility of these rankings and your magazine itself.Universities change very slowly - in many ways more slowly than even I would like.Yet,the people behind the U.S.News rankings lead readers to believe either that university quality pops up and down like politicians in polls,or that last year's rankings were wrong but this year's are right (until,of course,next year's prove them wrong).What else is one to make of Harvard's being #1 one year and #3 the next,or Northwestern's leaping in a single bound from #13 to #9?And it is not just this year.Could Johns Hopkins be the 22nd best national university two years ago,the 10th best last year,and the 15th best this year?Which is correct,that Columbia is #9 (two years ago),#15 (last year) or #11 (this year)?
Knowing that universities - and,in most cases,the statistics they submit - change little from one year to the next,I can only conclude that what are changing are the formulas the magazine's number massagers employ.And,indeed,there is marked evidence of that this year.
In the category "Faculty resources," even though few of us had significant changes in our faculty or student numbers,our class sizes,or our finances,the rankings' producers created a mad scramble in rank order,for example:
Down Last year This year Up Last year This year
Harvard #1 #11 MIT #6 #2
Stanford 3 15 Duke 13 4
Brown 12 22 Yale 10 6
Johns Hopkins 15 19
Dartmouth 18 24
One component of this category,"Student/faculty ratio," changed equally sharply,and not just in rank order but in what the magazine has presented as absolute numbers.Again,this is with very little change in our student or faculty counts:
Worse Last year This year Better Last year This year
Johns Hopkins 7/1 14/1 Chicago 13/1 7/1
Harvard 11/1 12/1 Penn 11/1 6/1
Stanford 12/1 13/1 Yale 11/1 9/1
Duke 12/1 14/1
Then there is "Financial resources," where Stanford dropped from #6 to #9,Harvard from #5 to #7.Our resources did not fall; did other institutions' rise so sharply?
I infer that,in each case,the formulas were simply changed,with notification to no one,not even your readers,who are left to assume that some schools have suddenly soared,others precipitously plummeted.
One place where a change was made openly was,perhaps,the most openly absurd.This is the new category "Value added." I quote the magazine:
Researchers have long sought ways to measure the educational value added by individual colleges.We believe that we have created such an indicator.Developed in consultation with academic experts,it focuses on the difference between a school's predicted graduation rate - based upon the median or average SAT or ACT scores of its students and its educational expenditures per student - and its actual graduation rate.
This passage is correct that such a measure has long been sought.However,like the Holy Grail,no one has found it,certainly not the "we" of this passage.The method employed here is,indeed,the apotheosis of the errors of the creators of these ratings:valid questions are answered with invalid formulas and numbers.
Let me examine an example in "Value added":The California Institute of Technology offers a rigorous and demanding curriculum that undeniably adds great value to its students.Yet,Caltech is crucified for having a "predicted" graduation rate of 99% and an actual graduation rate of 85%.Did it ever occur to the people who created this "measure" that many students do not graduate from Caltech precisely because they find Caltech too rigorous and demanding - that is,adding too much value - for them?Caltech could easily meet the "predicted" graduation rate of 99% by offering a cream-puff curriculum and automatic A's.Would that be adding value?How can the people who came up with this formula defend graduation rate as a measure of value added?And even if they could,precisely how do they manage to combine test scores and "education expenditures" - itself a suspect statistic - to predict a graduation rate?
Were U.S.News,under your leadership,to walk away from these misleading rankings,it would be a powerful display of common sense.I fear,however,that these rankings and their byproducts have become too attention-catching for that to happen.
Could there not,though,at least be a move toward greater honesty with,and service to,your readers by moving away from the false precision?Could you not do away with rank ordering and overall scores,thus admitting that the method is not nearly that precise and that the difference between #1 and #2 - indeed,between #1 and #10 - may be statistically insignificant?Could you not,instead of tinkering to "perfect" the weightings and formulas,question the basic premise?Could you not admit that quality may not be truly quantifiable,and that some of the data you use are not even truly available (e.g.,many high schools do not report whether their graduates are in the top 10% of their class)?
Parents are confused and looking for guidance on the best choice for their particular child and the best investment of their hard-earned money.Your demonstrated record gives me hope that you can begin to lead the way away from football-ranking mentality and toward helping to inform,rather than mislead,your readers.
The life and study between high school and college are completely different.
First, the study. one has half day free every week,and every day one gets up very early and stays up late at night in the high school.however in the college We nearly have four work days every week,and we have much free times,but this does not mean that the university life is boring and time wasting. In university one has some other very important things to do besides study, which plays a very crucial role in the college life.also the college study is not so boring. it covers a winder range in study which is not involved in the high school study.
Second,the life.people often says that university is like a small society which is accurate . one can touch and learn interpersonal relationship here.The university also has many activities and oppotunities and stages for one to show himself. what one thinks everyday is no longer only the study and scores but also how to get along with your roomates , classmates,teachers and other friends.College life is much more colourful and richful. at one's restime one can play basketball and pingpang and so on.
in a word, college is a small society and it is the first step you enter into socity and it is excellent and full of challenge. just enjoy it!7a686964616fe59b9ee7ad94364
e799bee5baa6e79fa5e98193e58685e5aeb9334more problem Interracial marriages can cause many problems within the family. Due to the fact that the couple has a different family background, culture, and custom, and social class level, many disagreements can occur. Religion can be a problem. If the two has a different religion then the other one, there is a problem. Because some religions celebrate certain holidays while others don't. In Christianity, for example, Christians celebrate Christmas while Buddudists don't. Eating habits can also be a problem. One might eat certain food the other don't or dislike. After all, their offspring will encounter problems too. Childrens who are mixed are not likely to be accepted in certain areas in a community. They often have a hard time fitting in. Interracial marriage is not a good idea, but it is still alright to marry someone who is different than you.