How might I get around the problem of formulas in a large worksheet taking ages to calculate or causing Excel to hang?
I have a large worksheet (300,000 lines x 50 columns) of horse racing data. The data are arranged one line per horse per race with column A holding the race identifiers. I need to normalise the data within each race on a scale from 0 to 1. For small amounts of data I use the formula x-min(r)/max(r)-min(r) in an array formula (where x= the item of data and r =the same data for the whole race).
My problem is, for such a large amount of data, the worksheet just hangs. For a smaller subset (5,000 rows) I switched off auto calc, copied the formulas as normal formulas and then converted them to array formulas. Even this took several hours. Once calculated the data can be saved as csv, the formulas don't need to be recalculated. Should I be looking to do this programatically rather than copying down the rows?
I have a large worksheet (300,000 lines x 50 columns) of horse racing data. The data are arranged one line per horse per race with column A holding the race identifiers. I need to normalise the data within each race on a scale from 0 to 1. For small amounts of data I use the formula x-min(r)/max(r)-min(r) in an array formula (where x= the item of data and r =the same data for the whole race).
My problem is, for such a large amount of data, the worksheet just hangs. For a smaller subset (5,000 rows) I switched off auto calc, copied the formulas as normal formulas and then converted them to array formulas. Even this took several hours. Once calculated the data can be saved as csv, the formulas don't need to be recalculated. Should I be looking to do this programatically rather than copying down the rows?