Hi,
I have spent the last couple hours compiling a large amount of data. I have a list of websites, each have a number of visits and transactions. The list is many thousands strong so i have a huge amount of data to work with but there are a lot of duplicate sites. See below, I have taken a snapshot of some of it:
I've highlighted a duplicate example.
Would it be possible across the entire data set to sum all of the duplicate rows into one. So, for the 3 shown, to be reduced to one row with 228 visits and 2 transactions?
And to do it quickly? As I could go through them manually but I don't have to time to go through 10,000 rows.
Thanks for the time and hope there is help out there, really appreciate it! Please let me know if what I'm asking isn't clear I can skype/email & explain further.
Thanks again!
I have spent the last couple hours compiling a large amount of data. I have a list of websites, each have a number of visits and transactions. The list is many thousands strong so i have a huge amount of data to work with but there are a lot of duplicate sites. See below, I have taken a snapshot of some of it:
I've highlighted a duplicate example.
Would it be possible across the entire data set to sum all of the duplicate rows into one. So, for the 3 shown, to be reduced to one row with 228 visits and 2 transactions?
And to do it quickly? As I could go through them manually but I don't have to time to go through 10,000 rows.
Thanks for the time and hope there is help out there, really appreciate it! Please let me know if what I'm asking isn't clear I can skype/email & explain further.
Thanks again!