

The following is a snapshot of how the data we are targeting to web scrap is placed in Wikipedia. I used Wikipedia for corona virus data and World bank website for tuberculosis and population data points. Wikipedia and World bank are the two websites I used for my project. While there were many data sources online that provided readily available corona virus datasets for analysis, I was more interested to get my hands dirty and fetch everything from the scratch just for the fun of it. The idea is to understand how tuberculosis and corona cases compare with each other. It may be because my awareness is limited, but then I decided to check it out with data. With all the buzz on the Corona virus cases, and how the world is struggling to address it, at a point of time fear changed into frustration, and I started wondering about other major infections, and why i could not remember the world impacted to this extent by a virus. It is also the reason I started working on this project. Web Scraping : Extracting, Cleaning and Organizing data with PythonĪ brief Data Prep project that uses Python to extract Corona virus and Tuberculosis data from web sources and make it ready for analysis BackgroundĬuriosity, is a trait that is driving a person’s or an organisation’s growth in today’s world.
