@Pallavi1 wrote:
Hi,
I would like to build a fraud predictive model for Financial services(fraud rate -approx.0.1%) to predict which business entity will have fraud in next 24 hrs. I have 100 million transactions for only 3 months data but my server can not process more than 20 million transactions.
I can not use 1 year worth data but even if i try to use only 3 months data, i can not process it. If the data is too big, will sampling work effectively? If yes, what are the best sampling techniques in this situation?
Posts: 1
Participants: 1