Deep Reinforcement Learning for Credit Card Fraud Detection
Lecture Notes in Networks and Systems, ISSN: 2367-3389, Vol: 891, Page: 285-297
2024
- 3Captures
Metric Options: CountsSelecting the 1-year or 3-year option will change the metrics count to percentiles, illustrating how an article or review compares to other articles or reviews within the selected time period in the same journal. Selecting the 1-year option compares the metrics against other articles/reviews that were also published in the same calendar year. Selecting the 3-year option compares the metrics against other articles/reviews that were also published in the same calendar year plus the two years prior.
Example: if you select the 1-year option for an article published in 2019 and a metric category shows 90%, that means that the article or review is performing better than 90% of the other articles/reviews published in that journal in 2019. If you select the 3-year option for the same article published in 2019 and the metric category shows 90%, that means that the article or review is performing better than 90% of the other articles/reviews published in that journal in 2019, 2018 and 2017.
Citation Benchmarking is provided by Scopus and SciVal and is different from the metrics context provided by PlumX Metrics.
Example: if you select the 1-year option for an article published in 2019 and a metric category shows 90%, that means that the article or review is performing better than 90% of the other articles/reviews published in that journal in 2019. If you select the 3-year option for the same article published in 2019 and the metric category shows 90%, that means that the article or review is performing better than 90% of the other articles/reviews published in that journal in 2019, 2018 and 2017.
Citation Benchmarking is provided by Scopus and SciVal and is different from the metrics context provided by PlumX Metrics.
Metrics Details
- Captures3
- Readers3
Conference Paper Description
Due to rapid advancement in electronic commerce technology, the use of credit cards has dramatically increased. As credit card becomes the most popular mode of payment for both online as well as regular purchase, cases of fraud associated with them are also rising. For many years, countless supervised machine learning approaches for anomaly detection have been proposed that have achieved state-of-the-art performance. For our deep reinforcement learning agent, we describe a revolutionary deep Q-network architecture and a customized OpenAI Gym environment in this paper. This design makes use of experience reply and is based on value function approximation. In order to carry out classification procedures determined by batches of input data; the deep Q-agent makes use of an epsilon-greedy policy. Subsequently, the agent is evaluated and rewarded by the OpenAI environment based on its performance in accordance with the assessment. The recollection of the agent contains every single detail of this adventure. The deep Q-agent takes several memory samples from its experience buffer at the end of the batch completing process. It then uses the Q-network to update the Q-value and the loss is calculated. Further, the weights are improved using backpropagation. The proposed procedure has achieved a level of performance that is state-of-the-art, and the results reveal that it was effective in correctly identifying fraudulent and non-fraudulent transactions.
Bibliographic Details
http://www.scopus.com/inward/record.url?partnerID=HzOxMe3b&scp=85187662776&origin=inward; http://dx.doi.org/10.1007/978-981-99-9524-0_22; https://link.springer.com/10.1007/978-981-99-9524-0_22; https://dx.doi.org/10.1007/978-981-99-9524-0_22; https://link.springer.com/chapter/10.1007/978-981-99-9524-0_22
Springer Science and Business Media LLC
Provide Feedback
Have ideas for a new metric? Would you like to see something else here?Let us know