Cookies help us display personalized product recommendations and ensure you have great shopping experience.

By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
SmartData CollectiveSmartData Collective
  • Analytics
    AnalyticsShow More
    composable analytics
    How Composable Analytics Unlocks Modular Agility for Data Teams
    9 Min Read
    data mining to find the right poly bag makers
    Using Data Analytics to Choose the Best Poly Mailer Bags
    12 Min Read
    data analytics for pharmacy trends
    How Data Analytics Is Tracking Trends in the Pharmacy Industry
    5 Min Read
    car expense data analytics
    Data Analytics for Smarter Vehicle Expense Management
    10 Min Read
    image fx (60)
    Data Analytics Driving the Modern E-commerce Warehouse
    13 Min Read
  • Big Data
  • BI
  • Exclusive
  • IT
  • Marketing
  • Software
Search
© 2008-25 SmartData Collective. All Rights Reserved.
Reading: Physicists, models, and the credit crisis, ctd.
Share
Notification
Font ResizerAa
SmartData CollectiveSmartData Collective
Font ResizerAa
Search
  • About
  • Help
  • Privacy
Follow US
© 2008-23 SmartData Collective. All Rights Reserved.
SmartData Collective > Big Data > Data Mining > Physicists, models, and the credit crisis, ctd.
Data MiningPredictive Analytics

Physicists, models, and the credit crisis, ctd.

DavidMSmith
DavidMSmith
5 Min Read
SHARE

Allen Engelhardt, physicist and former quant, provides a thoughtful response to my post the other day, where I asked, “Do physicists and engineers get similar [statistical] training?”

“Of course we do”, says Allen. He makes the point, though, that statisticians and physicists have different world-views.  A physicist will “(first) try to understand the model … He will run thought experiments. Then he might run some statistical tests against the data.”  I agree that’s a different process than that for a statistician, where in practice the data often come first and then, after data cleaning and exploratory analysis, a model is refined. 

I’d go so far as to say that even the word “model” means different things in the two disciplines.  In physics a model is a “true” representation of a reality, invariant and concrete … at least until some errant data point is discovered and a new model can be proposed, tested, and accepted. I use “true” there in quotes to reflect that scientific process: Newton’s Three Laws were “true” reflections of the physical world until relativity called for a new “true” model.  For a statistician in most disciplines, though, a model is merely a useful summariza…

More Read

First Look – New Wisdom RuleGuide
3 Organizations That Can See the Future with Predictive Analytics
Role Of Predictive Analytics In The Shifting Email Threat Landscape
Workshop on graphics in R by author of ggplot, Washington DC, 30-31 July
Resolving Many-to-Many Relationships

Allen Engelhardt, physicist and former quant, provides a thoughtful response to my post the other day, where I asked, “Do physicists and engineers get similar [statistical] training?”

“Of course we do”, says Allen. He makes the point, though, that statisticians and physicists have different world-views.  A physicist will “(first) try to understand the model … He will run thought experiments. Then he might run some statistical tests against the data.”  I agree that’s a different process than that for a statistician, where in practice the data often come first and then, after data cleaning and exploratory analysis, a model is refined. 

I’d go so far as to say that even the word “model” means different things in the two disciplines.  In physics a model is a “true” representation of a reality, invariant and concrete … at least until some errant data point is discovered and a new model can be proposed, tested, and accepted. I use “true” there in quotes to reflect that scientific process: Newton’s Three Laws were “true” reflections of the physical world until relativity called for a new “true” model.  For a statistician in most disciplines, though, a model is merely a useful summarization of noisy data. It reflects not an invariant truth about the underlying process that generated data, but a tool for identifying important effects and, sometimes, to make predictions.

I think we’re both in agreement that the problem isn’t the models themselves, or even the estimates from the models, but how those estimates were used and — crucially — by whom. Says Allen: “We knew and understood that the models were not valid on the tails, but there was no volume of trading on the tails, so it wasn’t very interesting.” That’s fine and dandy for the high-volatility trading group, where only consequences were the daily ups and downs within that one group. The danger came only when that model became a component of a VaR statistic reported to the upper management from the risk group, when the nuance of “not valid in the tails” was lost when it came to assigning capital allocation.

Allen goes on to say “Society is often expected to pick up the bill for tail effects… the cost of prevention may be bigger than the cost of fixing it.” Perhaps. But I certainly hope it remains in a financial institutions selfish self-interest (and I mean that only in a positive way) to avoid bankruptcy. This takes us into a longer discussion about the conflict between an individual banker’s short-term self-interest, and the long-term best interests of a bank. (I think Surowiecki had a column on this topic recently, but I can’t find it now.). But if there’s any lesson to be learned here, I hope it results in practices that lead these banks to understand the limitations of their modeling practices so that such failures can be avoided next time.

TAGGED:modeling
Share This Article
Facebook Pinterest LinkedIn
Share

Follow us on Facebook

Latest News

composable analytics
How Composable Analytics Unlocks Modular Agility for Data Teams
Analytics Big Data Exclusive
fintech startups
Why Fintech Start-Ups Struggle To Secure The Funding They Need
Infographic News
edge networks in manufacturing
Edge Infrastructure Strategies for Data-Driven Manufacturers
Big Data Exclusive
data mining to find the right poly bag makers
Using Data Analytics to Choose the Best Poly Mailer Bags
Analytics Big Data Exclusive

Stay Connected

1.2kFollowersLike
33.7kFollowersFollow
222FollowersPin

You Might also Like

PAW: Five Ways to Lower Costs with Predictive Analytics

6 Min Read

PAW: New Challenges for Developing Predictive Analytics Solutions

7 Min Read

Optimizing customer service levels with predictive analytics

7 Min Read
Image
Business IntelligenceCRMMarket ResearchSocial DataSocial Media AnalyticsUnstructured Data

Big Data Social Intelligence: Five Reasons Corporations Need It

10 Min Read

SmartData Collective is one of the largest & trusted community covering technical content about Big Data, BI, Cloud, Analytics, Artificial Intelligence, IoT & more.

data-driven web design
5 Great Tips for Using Data Analytics for Website UX
Big Data
ai is improving the safety of cars
From Bolts to Bots: How AI Is Fortifying the Automotive Industry
Artificial Intelligence

Quick Link

  • About
  • Contact
  • Privacy
Follow US
© 2008-25 SmartData Collective. All Rights Reserved.
Go to mobile version
Welcome Back!

Sign in to your account

Username or Email Address
Password

Lost your password?