Close Menu
    Trending
    • What If Your Portfolio Could Speak for You? | by Lusha Wang | Jun, 2025
    • High Paying, Six Figure Jobs For Recent Graduates: Report
    • What If I had AI in 2018: Rent the Runway Fulfillment Center Optimization
    • YouBot: Understanding YouTube Comments and Chatting Intelligently — An Engineer’s Perspective | by Sercan Teyhani | Jun, 2025
    • Inspiring Quotes From Brian Wilson of The Beach Boys
    • AI Is Not a Black Box (Relatively Speaking)
    • From Accidents to Actuarial Accuracy: The Role of Assumption Validation in Insurance Claim Amount Prediction Using Linear Regression | by Ved Prakash | Jun, 2025
    • I Wish Every Entrepreneur Had a Dad Like Mine — Here’s Why
    Finance StarGate
    • Home
    • Artificial Intelligence
    • AI Technology
    • Data Science
    • Machine Learning
    • Finance
    • Passive Income
    Finance StarGate
    Home»Machine Learning»Curse of Dimensionality. Before diving into my recent posts… | by Sudeep Chavare | Mar, 2025
    Machine Learning

    Curse of Dimensionality. Before diving into my recent posts… | by Sudeep Chavare | Mar, 2025

    FinanceStarGateBy FinanceStarGateMarch 15, 2025No Comments3 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    Earlier than diving into my latest posts about dimensionality discount methods, I ought to have ideally began right here — with the idea of the “Curse of Dimensionality.” At the moment, I’m studying the unbelievable e-book “Why Machines Study” by Anil Ananthaswamy, and I actually can’t suggest it sufficient! An in depth evaluation of this insightful e-book is certainly on its method, however I used to be so excited by the clear clarification of the Curse of Dimensionality that I couldn’t wait to share it.

    All of the insights I’m presenting listed here are closely drawn from Ananthaswamy’s glorious clarification, and the credit score absolutely belongs to him for simplifying such a fancy matter.

    There are a number of intuitive methods to know this “curse.” Let’s begin with a easy situation. Think about a dataset described by a single characteristic that ranges uniformly between 0 and a pair of, which means any worth inside this vary is equally doubtless. If we take twenty samples from this distribution, we’d see a roughly even unfold throughout the vary from 0 to 2.

    Now, suppose we have now two options, every independently ranging uniformly between 0 and a pair of. If we once more pattern twenty factors, however now from a 2D house (a 2×2 sq.), we’d discover one thing completely different. The area the place each options fall between 0 and 1 (the unit sq.) now covers solely 1 / 4 of our whole house. Thus, we would discover fewer factors — maybe solely 4 out of the unique 20 — on this area.

    Extending this to 3 options, every uniformly distributed between 0 and a pair of, we now have a cubic quantity. The unit dice (the place every characteristic is between 0 and 1) occupies solely an eighth of the overall quantity. If we once more pattern twenty factors, we would find yourself with even fewer — maybe simply two factors — inside our unit dice.

    This illustrates a basic problem in machine studying: because the variety of dimensions (options) will increase, the amount of house we should contemplate grows exponentially, inflicting knowledge factors to change into more and more sparse. With hundreds or tens of hundreds of dimensions, the prospect of discovering knowledge factors shut to one another turns into almost unattainable except the variety of samples is extremely massive. As Julie Delon from Université Paris–Descartes humorously places it, “In high-dimensional areas, no person can hear you scream.”

    Rising the variety of knowledge factors would possibly alleviate the issue, however this can also change into impractical in a short time. This problem is exactly why dimensionality discount methods change into invaluable. Strategies like Principal Part Evaluation (PCA), t-distributed Stochastic Neighbor Embedding (t-SNE), and autoencoders assist us simplify high-dimensional knowledge by figuring out and retaining probably the most informative options. By decreasing dimensions, these methods not solely mitigate the problems brought on by knowledge sparsity but in addition enhance computational effectivity, enabling machine studying algorithms to carry out higher. Primarily, dimensionality discount permits us to raised navigate the complexities of high-dimensional areas, serving to algorithms to listen to these essential indicators hidden amidst the noise.

    Reference : All the photographs included listed here are borrowed from Anil Ananthaswamy’s e-book, “Why Machines Study.”



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleLooking for Remote Work? Survey Says Remote Jobs Are Declining
    Next Article Mastering Hadoop, Part 3: Hadoop Ecosystem: Get the most out of your cluster
    FinanceStarGate

    Related Posts

    Machine Learning

    What If Your Portfolio Could Speak for You? | by Lusha Wang | Jun, 2025

    June 14, 2025
    Machine Learning

    YouBot: Understanding YouTube Comments and Chatting Intelligently — An Engineer’s Perspective | by Sercan Teyhani | Jun, 2025

    June 13, 2025
    Machine Learning

    From Accidents to Actuarial Accuracy: The Role of Assumption Validation in Insurance Claim Amount Prediction Using Linear Regression | by Ved Prakash | Jun, 2025

    June 13, 2025
    Add A Comment

    Comments are closed.

    Top Posts

    3 Lessons Entrepreneurs Can Learn from Frederick Douglass About Leading in Challenging Times

    February 1, 2025

    Why Every Small Business Owner Should Consider Real Estate — Even Without Deep Pockets

    June 10, 2025

    Why Every Company Should Have a 90-Day Cash Flow Buffer

    May 23, 2025

    OpenAI has upped its lobbying efforts nearly seven-fold

    February 1, 2025

    Adding Training Noise To Improve Detections In Transformers

    April 29, 2025
    Categories
    • AI Technology
    • Artificial Intelligence
    • Data Science
    • Finance
    • Machine Learning
    • Passive Income
    Most Popular

    Roadmap to Mastering Agentic AI. Agentic AI is rapidly transforming the… | by Kumar Nishant | Mar, 2025

    March 19, 2025

    Make Your Data Move: Creating Animations in Python for Science and Machine Learning

    May 7, 2025

    AI Is Taking Over Coding at Microsoft, Google, and Meta

    May 1, 2025
    Our Picks

    How Ending Penny Production Affects Consumers and Businesses

    May 23, 2025

    Reinventing Monopoly with Hierarchical Reinforcement Learning: Building a Smarter Game (Part 1) | by Srinivasan Sridhar | Mar, 2025

    March 7, 2025

    Citigroup Gives Employees Two Weeks of Remote Work in August

    June 11, 2025
    Categories
    • AI Technology
    • Artificial Intelligence
    • Data Science
    • Finance
    • Machine Learning
    • Passive Income
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 Financestargate.com All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.