Introduction
Fraudsters are constantly innovating—exploiting human psychology, technological loopholes, and global networks to steal billions. For banks,
fighting financial crime is an AI arms race against increasingly
sophisticated scams. By examining the playbook of scammers (from phishing
and Ponzi schemes to deepfakes and money laundering rings), banking
professionals can glean strategic insights to strengthen fraud detection, AML automation, customer experience, and innovation. Below, we
outline ten actionable lessons banks can ethically learn from the world’s
most cunning fraudsters, followed by a historical timeline of financial
fraud from the ancient world to the crypto era. Every claim is backed by
credible sources, and each lesson includes a real fraud example and a
banker’s takeaway.
Part 1: Ten Lessons from Fraudsters (for Bankers)
1. Master the Art of Social Engineering (for Good)
Fraud Example: Scammers are adept at social engineering—manipulating
people into bypassing security. A striking example occurred in 2019 when
criminals used an AI-generated deepfake voice to impersonate a CEO and
urgently direct a subordinate to wire $243,000 to a fraudulent account.
This tactic has evolved rapidly: in 2024, a Hong Kong-based firm was
scammed out of $25 million after an employee joined a video call with what
appeared to be familiar company executives—only to later discover they were
deepfake avatars controlled by fraudsters.
[1]
[2].
Similarly, Business Email Compromise (BEC) scams, where fraudsters pose as
trusted colleagues or vendors, have exploded. Between 2016 and 2019, BEC
scams cost organizations worldwide over $26 billion in losses
[3].
These attacks prey on human trust and fear—e.g. an “urgent” request from
the boss or a panic-inducing email from “IT support” prompting a password
reset.
Banker’s Takeaway: Banks should assume that anyone can be phished or impersonated . Just as scammers exploit human psychology, banks must fortify it.
Implement strict multi-factor verification for high-risk transactions (e.g.
call-back confirmations for large wires) and provide ongoing security
training to staff and customers. Internally, encourage a “trust but verify”
culture for unusual requests—even if seemingly from the CEO. Externally,
leverage AI-driven anomaly detection to spot telltale signs of account
takeover or social engineering. For example, customer behavior analytics can flag if a normally inactive account suddenly initiates rapid transfers,
prompting an intervention. By using advanced transaction analysis tools to monitor for out-of-pattern activities in real time
[4]
[5],
banks can catch socially engineered fraud before money walks out the door. Modern solutions now even employ AI to analyze transactional patterns or graph visualizations of fund flows to detect the subtle footprints of social engineeringin
action.
2. Create Urgency and Trust – Then Verify Reality
Fraud Example: Many scams succeed by crafting a compelling narrative
that builds trust and urgency. Consider classic Ponzi schemes: con
artists like Bernie Madoff spent years cultivating a reputation as a genius
investor and philanthropist, lulling victims (and even banks) into a false
sense of security. Madoff’s fund reported amazingly steady returns for
decades – too steady, in hindsight – and attracted $65 billion in
investor funds before collapsing in 2008
[6].
Early investors were paid “profits” with new investors’ money, a house of
cards sustained by confidence and the allure of exclusive, high-yield
returns. Another example: Allen Stanford, who ran a $7 billion Ponzi,
sponsored cricket tournaments and charitable events to appear legitimate
[7]
[8].
These schemes masterfully exploit greed, FOMO, and trust in authority (even King George endorsed the South Sea Company in 1720’s infamous bubble,
see Part 2). Victims often overlook red flags due to the scammer’s aura of
credibility and the urgency to not “miss out.”
Banker’s Takeaway: Fraudsters are storytellers – so banks must
pierce the narrative with data. When a client’s investment returns look
“too good to be true” or funds flow in unusual ways, don’t be blinded by
their prestige or longevity. Use objective verification and continuous
monitoring even for long-time, “trusted” customers. Madoff, for instance,
was a well-known figure, yet JPMorgan ignored mounting red flags; for years
Madoff laundered billions through a single bank because warnings were not
escalated
[9]
[10].
The lesson is to verify trust with transparency: implement strict KYC/KYB (Know Your Customer/Business) reviews and periodic
re-validation of customers’ activities against their profiles. Encourage a
healthy skepticism in compliance teams – if an entity’s transaction patterns
defy industry norms, investigate why. Banks can also deploy data visualization dashboards that cut through complex account histories to reveal anomalies (e.g. an account consistently yielding high returns regardless of market
conditions). An ethical twist on the scammer’s playbook is for banks to
create their own urgency around due diligence – promptly follow up on any
inconsistency or tip (whistleblowers warned about Madoff years earlier). In
short, trust your clients, but trust your data and controls more.
3. Exploit Data (Like Hackers Do) to Predict Fraud
Fraud Example: Modern fraudsters treat data as their treasure.
Through hacks and data breaches, they steal personal information to
impersonate customers or create synthetic identities. For example, synthetic identity fraud – where criminals combine real stolen data (like a Social Security number)
with fake details to build a new “person” – cost U.S. banks an estimated $20 billion in 2020
[11].
These fake personas opened accounts and credit lines that looked legitimate
until they “busted out” with unpaid loans. Fraud rings also harvest leaked
passwords to take over online banking accounts, or use malware to collect
card data. A single breach can fuel countless frauds: the 2017 Equifax
breach exposed data of 147 million Americans, leading to years of identity
theft attempts. Criminals even share tools and stolen data on the dark web,
essentially running fraud as an organized digital business.
Banker’s Takeaway: Banks should weaponize data for defense just as crooks do for offense. This means investing in advanced analytics, AI-driven pattern recognition, and information-sharing. First, break
down silos: integrate data from across the enterprise (transactions,
logins, customer profiles) to get a 360° view. Then, deploy machine
learning to detect subtle patterns – for instance, multiple accounts using
the same phone number or address (a clue to synthetic IDs), or a customer’s
device fingerprint suddenly changing. Transaction monitoring systems should flag if a usually local customer starts transacting from foreign IP
addresses or if funds begin routing through new intermediaries atypical for
that client. Modern AML systems can harness graph analytics to
connect seemingly unrelated accounts that frequently trade or funnel money
between each other. Such network analysis can unmask mule networks (one bank uncovered a ring of 90+ accounts linked
by a common email syntax). Additionally, banks must collaborate: share
threat intelligence on emerging fraud tactics (via industry groups or
schemes like the FBI’s Fusion Cells). By analyzing big data trends – much
like fraudsters do when scoping victims – banks can predict and preempt scams , rather than just react. Think of it as using the scammers’ data-driven
mindset against them, ethically: if hackers sift data to find
vulnerabilities, banks should sift data to find hackers.
4. Stay Agile and Adapt to New Threats
Fraud Example: If scammers have a superpower, it’s agility. They
pivot quickly to exploit news and technology. When the COVID-19 pandemic
hit, fraudsters launched a blitz of new schemes in weeks – from fake PPP
loan websites to unemployment benefit fraud on a massive scale. In the
U.S., over $100 billion was siphoned off through fraudulent pandemic
unemployment claims
[12]
[13]
as criminals took advantage of overwhelmed systems. Internationally,
criminals shifted tactics overnight (e.g. targeting remote workers with
phishing emails or spoofing health authorities). Likewise, as faster payment
systems roll out, scammers have adjusted: the U.K. saw a surge in
“authorized push payment” scams where victims are conned into sending
instant bank transfers, leading to £479 million in losses in 2020
(prompting new refund rules)
[14]
[15].
Fraudsters even iterate their scripts based on what works, A/B testing
phishing messages for higher click rates.
Banker’s Takeaway: Speed is survival. Banks must embrace an
agile, proactive approach to fraud and compliance – bureaucratic lag can be
fatal. This means shortening the feedback loop: when a new scam or modus
operandi is reported, rapidly update rules and train staff. For example, as
soon as COVID relief fraud patterns emerged, banks that quickly implemented
extra verification for out-of-state benefit deposits or flagged mismatches
in claimant names saw far fewer losses. Agile banks are also adopting
real-time fraud detection on payments (not waiting until after money is
gone). A real-time AI engine can decline or pause suspect
transactions mid-stream if risk signals hit (for instance, a sudden large
wire to a first-time beneficiary in a high-risk country)
[16]
[17].
Just as scammers tweak their tactics on the fly, banks should consider
“fraud SWAT teams” that monitor emerging threats and adjust controls
immediately (daily if needed). Regular red team exercises can help:
task internal teams or external experts to simulate new fraud techniques
against your systems to find gaps before real criminals do. An agile bank
also means empowering front-line employees – give branch tellers and call
center reps the authority to pause a transaction if something feels off
(many elderly customers have been saved from scam transfers by an alert
banker who intervened). In essence, outmaneuver criminals by moving as fast
as they do; anticipate trends (e.g. deepfake scams, crypto fraud) and be
ready with countermeasures before they go mainstream.
5. Follow the Money Trail – Criminals Do
Fraud Example: Criminal networks excel at obfuscating the money
trail. They break transactions into many hops and layers, often using
“money mules” – accomplices (witting or unwitting) who receive and forward
illicit funds to hide the true beneficiary. For instance, European
authorities uncovered mule networks where hundreds of individuals were
moving illicit proceeds through small transfers, adding layers of distance
between the criminal and the cash
[18]
[19].
Organized rings also leverage shell companies and intermediaries to launder
money across borders (a notorious example: the Russian “Mirror Network”
laundered billions through a daisy-chain of banks and offshore firms).
Fraudsters essentially map out financial networks and exploit gaps in
inter-agency or inter-bank communication. The 2016 Bangladesh Bank
cyber-heist is a case in point: hackers sent $81 million in fraudulent
transfers via the SWIFT network, then funneled the money through casinos
and money changers in the Philippines – a jurisdictional maze that
complicated recovery
[20]
[21].
Banker’s Takeaway: Think like a forensic accountant – map the
entire web of transactions, not just individual nodes. Banks can invest in
tools that build graphical analysis of accounts and transactions, illuminating hidden connections (like shared
IP addresses, common beneficiaries, or cyclic money flows) that manual
reviews miss. By visualizing the “money map,” compliance teams can follow
suspicious funds through complex paths and uncover the bigger picture of
organized fraud. Importantly, share information: fraudsters count on
banks not talking to each other. Use legal gateways (e.g. the USA PATRIOT
Act Section 314(b) or Europe’s FIU cooperation mechanisms) to share
intelligence on suspected mule accounts or emerging laundering patterns. If
one bank sees a customer receive multiple small deposits and then immediately wire them out (classic mule behavior), alerting other
institutions can prevent the next step in the chain. Also, coordinate with
law enforcement early; even partial data (like a mule’s name or a shell
company’s registration) can help paint the full picture when combined with
intelligence from other sources. In short, networked crime demands a networked defense . Fraudsters leverage global financial connectivity – banks must do the
same, partnering across institutions and borders to trace illicit flows. By
mastering “follow-the-money” techniques, banks turn the tables: complex
laundering schemes become high-resolution maps pointing to the
perpetrators.
6. Embrace Technology – Because Fraudsters Will
Fraud Example: The most successful scammers are often tech
early-adopters. They were among the first to leverage AI for crime –
from deepfake audio (as in the CEO phone scam) to AI-written phishing
emails that evade spam filters. Criminal rings use automated botnets to
test stolen card numbers or credentials at scale, and encrypt their
communications on the dark web. In recent years, criminals have exploited
cryptocurrencies and blockchain tech: mixers like Tornado Cash were
used to launder over $7 billion in dirty crypto since 2019
[22]
by algorithmically breaking and rejoining transaction trails. Even deepfake
videos have emerged – in one 2023 case, fraudsters used a deepfake of a
company director on a Zoom call to authorize a transfer (the attendees had
no idea they were speaking with an AI impostor). On the flip side, banks
have started deploying AI in defense – but the race is tight. As one
fintech risk expert described, it’s a constant “digital arms race” with
each side trying to outpace the other
[23]
[24].
Banker’s Takeaway: Leverage good AI to fight bad AI. Banks
should actively incorporate emerging technologies (AI, biometrics,
blockchain analytics) into their anti-fraud arsenal as fast as – or faster
than – criminals do. For instance, use AI-powered voice recognition to
detect if a caller’s voice matches the customer’s voiceprint on record,
which could thwart deepfake phone fraud. Similarly, deploy machine learning
models that continuously learn from new fraud patterns; unlike static
rules, adaptive AI can catch novel schemes that don’t match yesterday’s
profiles
[25]
[26].
One powerful approach is link analysis : AI can sift through millions of transactions to find patterns (say, a
cluster of accounts transacting in round amounts at odd hours) that humans
miss, boosting proactive fraud detection. Banks are also starting to use computer vision on IDs and selfies to spot deepfake or forged documents
during onboarding (there are AI solutions now that detect if an uploaded
“photo ID” is actually a digitally manipulated image). On the
cryptocurrency front, embrace blockchain forensic tools – if criminals use
crypto, banks can trace crypto. Blockchain analysis firms can flag
addresses linked to illicit activity, enabling banks to screen and even
freeze transfers to mixers or sanction-listed wallets. The key is a mindset
shift: innovate in defense. Just as scammers aren’t afraid to
beta-test new tech to steal, banks should pilot cutting-edge tech to
secure. Allocate R&D budget for compliance innovation, encourage
fintech partnerships, and cultivate an internal culture that is excited
about new tools (rather than wary of them). In summary, don’t bring a knife
to an AI gunfight – arm up with the latest technology for a fighting chance
in the fraud wars.
7. Red-Team Your Own Bank (Think Like a Criminal)
Fraud Example: Ever notice how scammers seem to know exactly where
the weak spots are? They probe constantly for vulnerabilities – whether
it’s a bank website’s less-secure subdomain, an employee’s publicly exposed
email, or a loophole in transaction screening thresholds. A notorious
example is the 2016 Bangladesh Bank heist: attackers gained entry likely
through an unpatched system, then discovered the bank’s printer used for
SWIFT confirmation was off on weekends (helping delay detection)
[27]
[28].
They even cleverly misspelled “foundation” as “fandation” in a transfer
instruction to evade a watchlist trigger (the word “foundation” was flagged
at the Fed, but “fandation” passed). This kind of insider knowledge comes
from thinking like an attacker. Fraudsters often do dry runs – sending
small test transactions to see if they get flagged, or calling customer
service with partial info to see what verification is asked. Essentially,
they red-team our financial systems relentlessly.
Banker’s Takeaway: It’s time to beat them at their own game: conduct regular adversarial testing on your bank . Banks can hire professional ethical hackers or form internal red teams to
simulate fraud attempts and social engineering attacks. For example, test
your new account onboarding by attempting to open accounts with synthetic
identities – do the fake IDs get caught? Launch phishing email tests at
employees to gauge who clicks, then train and repeat. Try “smishing” (SMS
phishing) or vishing (voice phishing) your own call centers to see if
agents properly verify caller identity. In IT, run penetration tests not
just for data breaches but specifically to see if someone could alter
transaction data or exploit a business process (e.g. is there a way to
spoof an email approval for a wire transfer?). The findings from these
exercises should directly inform stronger controls. Perhaps your red team
finds they could walk into a branch with a forged driver’s license and
withdraw $5,000 from a random account – use that to enhance teller training
and authentication steps. Or maybe a tester finds your wire room will
process a faxed request with a mere signature copy – time to require
call-backs for faxed instructions. This process is essentially quality assurance through a criminal lens . Regulators are increasingly supportive of such simulated attack
exercises, and some jurisdictions even mandate them for cyber resilience.
The more you think like a thief, the more secure your institution
becomes. After all, who better to reveal your blind spots than someone
creatively emulating the bad guys? Find the holes and patch them before a real criminal does.
8. Leverage Insider Knowledge and Collaboration
Fraud Example: While we often picture outside hackers, many frauds
involve insiders or at least insider knowledge. Criminals skillfully
recruit bank insiders (for example, a helpdesk employee bribed to give up a
one-time password, or a loan officer colluding to approve bogus loans). In
other cases, criminals benefit from the lack of collaboration among
legitimate parties. A famous failure of communication was the case of Bernie Madoff again – different departments and banks had suspicions, but no one
connected the dots until it was too late. JPMorgan’s London desk noticed
odd transactions, and its U.S. compliance team had concerns, yet these were
not effectively shared internally or with regulators
[29]
[30].
Fraudsters thrive in these silos. Similarly, cross-border scams exploit the
fact that Bank A in country X doesn’t talk to Bank B in country Y – by the
time authorities piece things together, the money (and criminals) are long
gone.
Banker’s Takeaway: Break down the silos – within your bank and
across the industry. Internally, ensure that fraud intelligence is shared
enterprise-wide. If your credit card division sees a spike in fraud rings,
alert the deposit account teams who might soon see related activity.
Establish an internal fraud knowledge hub and regular cross-department
briefings. Technology can help: use case management systems that different
units (fraud, AML, cybersecurity) can all contribute to and review. This
way, a suspicious pattern in one area raises a company-wide alarm.
Externally, participate in information-sharing forums. Many countries have
set up banker alliances against fraud (like the UK’s JMLIT – Joint
Money Laundering Intelligence Taskforce – which pools intel between banks
and law enforcement). In the U.S., banks can share customer fraud
information under safe harbor via Section 314(b) of the PATRIOT Act. Taking
advantage of these can turn isolated insights into collective action. Insider fraud risk also means engaging and monitoring your own employees. Cultivate a
culture where employees feel accountable to report unethical behavior, and
rotate staff in sensitive roles to prevent long-term collusion. Conduct
proactive screenings (within legal limits) to detect if staff might be under
financial duress or exhibiting risky behavior (many banks use behavioral
analytics on employees’ system access patterns to spot possible rogue
actions). Ultimately, fraud fighting is a team sport. By collaborating and
sharing knowledge, banks deny fraudsters one of their key advantages: that
defenders are divided. When banks and agencies unite, the fraudulent
schemes that once slipped through the cracks can be spotted and stopped via
collective vigilance.
9. Layer Defenses to Counter Layered Crimes
Fraud Example: Fraudsters rarely rely on a single point of failure –
they layer their schemes to maximize success. Take money laundering:
criminals often use a layering phase, shuffling illicit funds through
multiple bank accounts, shell companies, and jurisdictions to muddy the
trail. One of the largest money laundering scandals in history involved
Danske Bank’s Estonian branch, where around €200 billion in
suspicious non-resident payments flowed through 2007–2015. The funds came
from Russia and elsewhere and were funneled through complex layers of shell companies to hide the real owners
[31]
[32].
Authorities later found that thousands of these shell company accounts had
fictional owners or opaque structures, making it easy for criminals
(potentially including sanctioned individuals) to move money unnoticed. Only
a combination of whistleblower reports and deep investigation uncovered the
full layering. Another example: shell firms were at the heart of the
Panama Papers leak (2016), which exposed how corrupt officials and tax
evaders used offshore entities to conceal wealth, prompting new beneficial
ownership transparency laws worldwide.
Banker’s Takeaway: To defeat layered fraud, build layered defenses . One layer is customer due diligence: strengthen your onboarding and KYC
processes to verify beneficial owners of companies. Don’t just collect
paperwork – use external data sources and databases to corroborate that
“John Doe” owning XYZ Corp isn’t a straw man. With modern APIs, banks can
cross-check corporate registries and sanction lists in real time for new
business accounts. Another layer is transaction monitoring that looks not just at one account in isolation, but across all accounts
for patterns (e.g. multiple business accounts all sending increments of
$9,900 to the same offshore entity – likely structuring to avoid reporting).
Implement risk-based trigger events: for instance, if an account
starts receiving or sending significantly more money than when originally
profiled, require a fresh review (dynamic KYC). Employ enhanced due diligence for high-risk corridors – Danske’s Estonian branch failure was partly due
to lax oversight of high-risk non-resident flows. Had they layered robust
compliance (like requiring originator identification on incoming wires,
checking shell company owners, and scrutinizing unusual volume spikes), the
scheme could have been flagged much earlier
[33]
[34].
Regulators responded by increasing banks’ capital requirements and
mandating reforms after that scandal
[35].
Banks should take the hint: anticipate regulators by self-imposing tougher
multi-layer controls. Think of defense in depth: even if one check misses a
clever fraud, another will catch it. A fraudulent wire might bypass an
automated rule, but then get picked up in a daily manual review of large
exposures. Or a shell company account might open successfully, but its first
big transaction triggers a freeze until verified. Each layer – KYC,
monitoring, analytics, audits, training – adds friction for the fraudster
and reduces the chance that a multi-stage scam can fully succeed.
10. Understand (and Counter) the Psychology of Fraud
Fraud Example: At the core, scams target the human element – emotions
and cognitive biases. Whether it’s fear, greed, compassion, or impatience,
fraudsters know how to push buttons. Phishing emails often create a
panic (“Your account will be closed in 24 hours!”) to prompt quick,
irrational action. Romance scams exploit loneliness and trust,
persuading victims over months that they’re in a loving relationship, only
to request money for a fake emergency. Tech support scams use fear
and confusion (“Your computer is infected! Call us immediately!”) to
extract payments. One infamous con is the “grandparent scam,” where
fraudsters call an elderly person pretending to be a grandchild in distress
needing cash bail – playing on both fear and love. The psychology of scarcity is also abused: consider NFT or crypto pump-and-dump schemes where
scammers hype a “limited opportunity” to create FOMO (fear of missing out),
driving victims to invest quickly. In short, fraudsters are amateur
psychologists, adept at getting people to suspend rational judgment.
Banker’s Takeaway: Banks should integrate behavioral science into fraud prevention and customer experience. Start by educating customers
about these psychological tactics – informed customers are less likely to
panic-click a phishing link or send money to a “stranded relative” without
verification. Many banks now run public awareness campaigns highlighting
common scam scripts, which has been shown to significantly reduce success
rates. Next, incorporate friction at points of emotional decision. For
example, if an elderly client attempts an unusually large online transfer,
a smart system might present a gentle warning: “Is this transfer related to
someone claiming to be your relative in an emergency? It might be a scam –
consider calling them first.” Some UK banks have implemented “confirmation
of payee” and even delay mechanisms for first-time payees, giving
fraud teams a window to contact the customer and verify high-risk transfers
(especially for vulnerable customers). On the flip side, improve legitimate
customer experience by learning from scammer communication techniques
(ethically!). Scammers excel at simple, persuasive messaging – banks can
likewise communicate security measures in clear, non-technical language
that resonates, rather than legalistic jargon that customers might ignore.
Additionally, use customer behavior analytics to detect when a
customer’s behavior seems off for them – perhaps indicating they are being
socially engineered. If a typically frugal customer suddenly is wiring
money overseas, a well-timed automated SMS or call to verify intent can
save them from fraud. Finally, ensure your customer-facing staff are
trained in empathy and scam recognition. Often a teller or call rep, sensing
confusion or distress, can tactfully ask questions that uncover a scam in
progress (“Mind if I ask what this $5,000 withdrawal is for? We often see
scammers pressure people to withdraw cash like this.”). In essence,
counteract the fraudster’s psychology with your own: use nudges, education,
and empathetic intervention to keep customers making decisions in their
best interest. Protecting people from fraud isn’t only about tech and rules
– it’s about understanding people.
Part 2: A Timeline of Financial Fraud – From Ancient Schemes to the Crypto and ChatGPT Era
Fraud is not new – it’s been with us through every era of finance, evolving
with technology and prompting new laws in response. Below is a
chronological timeline highlighting key periods in financial fraud history,
the enabling developments of each era, notable cases, and how authorities
responded.
300 B.C. – Ancient Origins of Fraud
Development: The concept of credit and insurance in antiquity created
the first opportunities for financial deceit.
Case: One of the earliest recorded financial frauds was perpetrated
by a Greek merchant named Hegestratos around 300 B.C. He took out an
insurance policy (a bottomry loan) on a ship and its cargo of corn –
agreeing to repay the loan with interest if the cargo arrived safely.
Hegestratos plotted to sink his empty ship and keep the loan money,
effectively cheating the lender. His plan failed when he was caught in the
act and drowned fleeing his crew
[36]
[37].
This foiled scheme stands as the first documented insurance fraud in
history.
Response: Ancient lawmakers soon recognized the need for maritime
trade regulations. Greek and Roman laws against fraud (such as the lex
Rhodia on maritime law) were instituted to punish such deceit. While
primitive by modern standards, these early legal codes laid the groundwork
that fraudulent contracts or sabotage for gain would not be tolerated.
1710s–1720s – The First Stock Bubbles and Scandals
Development: The birth of stock companies and public markets in the
early 18th century enabled mass speculation – and fraud. Governments
granted monopolies to enterprises, and unscrupulous insiders took advantage
of investor mania.
Case: The South Sea Bubble of 1720 in England stands as an
early colossal scam. The South Sea Company, granted a monopoly on trade
with Spanish America, was hyped with extravagant promises. Company
insiders and even government ministers engaged in fraudulent schemes – including bribery and insider trading – to pump the stock price despite
the company having meager real prospects
[38]
[39].
At its peak, shares soared over 800% (from £128 to £1,000), only to
collapse disastrously, losing 80% of value by year’s end
[40]
[41].
Countless investors were ruined, and investigations revealed widespread
corruption and false accounting by the company’s directors. (Notably, Sir
Isaac Newton lost a fortune and lamented he could “calculate the motions of
heavenly bodies, but not the madness of people.”)
Response: The British government was spurred to act. Parliament
convened inquiries that exposed the fraud and led to disgrace and
punishment for some South Sea Company officials. In 1720, in reaction to
the bubble, Parliament passed the Bubble Act, aiming to prevent
“fraudulent projects” by requiring royal approval for joint-stock
companies
[42]
[43].
Although the Bubble Act proved problematic and was repealed a few years
later, its intent was an early attempt at securities regulation. Public
anger from the South Sea fiasco also contributed to more cautious investor
attitudes and a lasting metaphor: “South Sea Bubble” became a byword for
financial fraud and folly. Across the English Channel, France experienced a
similar Mississippi Company bubble under John Law – which likewise burst
and led to economic turmoil – reinforcing the lesson that exuberant markets
need oversight to deter fraud.
1792 – The First Market Crash in America
Development: Post-Revolution America saw the emergence of government
bonds and the first stock exchanges. However, regulation was nonexistent,
and speculators with insider access could easily manipulate markets.
Case: William Duer, a former assistant Secretary of the
Treasury and a member of George Washington’s inner circle, engineered
America’s first Wall Street scandal. In 1792, Duer used his inside
knowledge of Treasury operations to speculate heavily in government bonds
and bank stocks, often using leverage and spreading rumors to sway prices
[44]
[45].
When his overleveraged bets collapsed, it triggered the Panic of 1792 – the young nation’s first financial crash. Duer’s downfall landed him in
debtor’s prison (where he died in 1799), and many of his investors were
wiped out
[46]
[47].
Response: The shock of 1792 had a silver lining: it led to greater
organization of the securities market. Just weeks after the crash, traders
in New York gathered under a buttonwood tree on Wall Street and signed the Buttonwood Agreement – effectively creating the NYSE – to establish
rules and commissions for fair trading
[48].
In essence, this was an agreement to restore trust and prevent gross
manipulation: brokers would only deal with each other, at fixed minimum
commissions, curbing some speculative excess. While it wasn’t government
regulation, it was the market self-regulating to avoid another Duer debacle.
This early episode taught American financiers the need for transparency and
limits on insider dealing – lessons that would echo in future crises.
19th Century – Counterfeits and Corporate Scandals
Development: The 1800s saw paper money, rapid industrialization, and
big finance – along with new fraud challenges. The absence of a national
bank in the U.S. and proliferation of state bank notes led to rampant
counterfeiting; later in the century, large corporate enterprises created
opportunities for accounting fraud and stock swindles.
Case (Counterfeiting): By the end of the U.S. Civil War (1865),
counterfeit currency was so prevalent that an estimated one-third of all U.S. money in circulation was fake
[49]
[50].
Gangs of counterfeiters took advantage of primitive printing controls and
the chaos of war to flood the economy with phony banknotes, threatening the
young nation’s financial stability.
Response: On the last day of the Civil War in 1865, President Lincoln
created the United States Secret Service – not for presidential
protection, but specifically to combat counterfeiting
[51].
The Secret Service cracked down on counterfeit rings and introduced
standardized, harder-to-forge currency designs. This dramatically reduced
fake notes and saved the post-war economy. Across the Atlantic, in
Victorian Britain, an era of railway booms and busts brought corporate
frauds – like the 1840s railway stock swindles that led to calls for
company law reform. By the late 19th century, governments started
instituting more robust corporate disclosure requirements (e.g. Britain’s
Joint Stock Companies Act) as a response to fraudulent stock promotions
that had fleeced investors. The creation of professional auditing also
traces back to this period as a defense against management fraud. In short,
the 19th century taught regulators that systemic fraud (be it counterfeit money or corporate deceit) can undermine entire economies, and that law enforcement and legal
frameworks must evolve in step.
1920s – Ponzi Schemes and the Great Crash
Development: The roaring 1920s featured stock market euphoria and
get-rich-quick schemes. New communication tools (telegraphs, phones, mass
mail) allowed frauds to reach more victims. Crucially, a lack of federal
securities laws meant caveat emptor – fraudsters could operate with
impunity until they collapsed.
Case: In 1920, Charles Ponzi became an American legend for
fraud. He promised investors in Boston a 50% profit in 90 days, claiming
arbitrage of international postal reply coupons. In reality, Ponzi was paying early investors with money from later ones – the blueprint of the “Ponzi scheme” that now bears his name. At its
height, Ponzi was raking in cash from about 40,000 people, with an
estimated $15 million in principal invested (over $190 million today) –
before it crashed. When authorities shut him down, investors lost about $20
million
[52].
The term “Ponzi scheme” entered the lexicon to describe this type of
pyramid fraud
[53].
Just years later, the 1929 stock market crash exposed massive market
manipulation: pools of speculators had secretly colluded to pump up stock
prices, then dump them on the public. Insider trading was rampant, and many
companies sold stock on false claims. The crash wiped out billions and
ushered in the Great Depression.
Response: The federal government responded with a regulatory
revolution. In the aftermath of 1929, the U.S. Congress passed the Securities Act of 1933 and the Securities Exchange Act of 1934, which, for the first time,
imposed rigorous requirements on securities offerings and trading. The 1934
Act created the U.S. Securities and Exchange Commission (SEC) to
enforce these rules
[54].
These laws mandated truthful disclosure in stock sales and outlawed
manipulative practices, fundamentally changing the game. The SEC’s
establishment in 1934 (explicitly “in response to the stock market crash of
1929”)
[55]
was a direct answer to the excessive fraud and speculation of the 1920s.
Additionally, Ponzi’s scam (and others like it in the 1920s) led many
states to pass “Blue Sky Laws” (securities regulations) even before the
federal acts, aiming to protect investors from “blue sky and hot air.” The
legacy of this era is clear: robust financial markets require transparency
and oversight to prevent fraud from again reaching systemic levels.
1970s–1980s – Money Laundering and Global Crackdown
Development: By the late 20th century, organized crime and drug
cartels were generating vast illegal profits, which they sought to launder
through banks. Financial globalization accelerated, and criminals took
advantage of weak links in international oversight. The 1970s also saw the
rise of offshore banking and secret jurisdictions.
Case: One landmark scandal was the Bank of Credit and Commerce International (BCCI) in the 1980s. This Luxembourg-chartered, Pakistani-run bank operated in 70
countries and became known as the “Bank of Crooks and Criminals
International.” BCCI engaged in massive fraud and money laundering—creating
fake loans, using shell companies, and secret books to hide
losses and criminal clients. When investigators finally cracked down, they
found a $5 to $10 billion hole in BCCI’s balance sheets. In 1991,
regulators shut down BCCI worldwide for massive fraud; the Bank of
England governor stated, “The culture of the bank is criminal”
[56]
[57].
BCCI had been laundering money for drug lords, arms traffickers, and
dictators, exploiting global jurisdictional blind spots.
Response: The world’s regulators realized that dirty money was a
global problem requiring global solutions. The United States had already
taken a step in 1970 with the Bank Secrecy Act (BSA) – which
required banks to keep records and report cash transactions over $10,000,
establishing the framework for modern anti-money laundering (AML) controls
[58].
In 1986, the U.S. made money laundering itself a federal crime (Money
Laundering Control Act). On the international stage, the Financial Action Task Force (FATF) was created by the G7 in 1989 specifically to combat money laundering on a
global scale
[59]
[60].
FATF issued 40 Recommendations that became the blueprint for AML laws
worldwide. After BCCI, many countries strengthened oversight of
international banks: e.g., higher due diligence for cross-border banking
and better cooperation among regulators. The European Community and others
also accelerated banking transparency directives in the 1990s. These
efforts started paying off: banks implemented stricter customer
identification, suspicious activity reporting (SAR) systems, and
cross-border information sharing through entities like the Egmont Group of
financial intelligence units. In sum, the late 20th century marks the era
when money laundering was recognized and attacked as a distinct crime , and banks were conscripted as frontline defenders through laws and global
standards. Financial institutions that once turned a blind eye to illicit
deposits were now required to “know their customer” or face heavy penalties
(as many learned in subsequent decades).
2008 – The $65 Billion Ponzi and Great Recession Frauds
Development: The 2000s saw complex financial engineering and a
housing boom, but also lapses in risk management and ethics. Fraud surfaced
in both high finance (e.g. mortgage-backed securities misrepresentations)
and low finance (predatory lending). As the decade ended in crisis, several
massive frauds came to light, shaking trust further.
Case: The most infamous modern Ponzi scheme was perpetrated by Bernard L. Madoff , a respected financier who ran a secret investment advisory business that
turned out to be a decades-long fraud. When Madoff’s pyramid collapsed in
December 2008, it revealed about $65 billion in fictitious profits and principal in client accounts – the largest Ponzi scheme ever
[61].
Madoff had promised steady ~10% annual returns and fabricated account
statements to clients, all while simply shuffling money. The fraud went
undetected by regulators (despite warnings) for years, exploiting gaps in
oversight between the SEC, FINRA, and international regulators. Around the
same time, another fraudster, Allen Stanford, was exposed for running an $8
billion Ponzi via bogus “high-yield” CDs at his Antigua-based bank. And
beyond Ponzi schemes, the 2008 financial crisis itself had elements
of fraud: banks faced accusations of mis-selling mortgage-backed securities
and a few executives (like at Bear Stearns hedge funds) were tried for
securities fraud (though acquitted). The crisis also uncovered mortgage
fraud on the consumer level – millions of falsified income statements
(“liar loans”) contributed to the housing bubble.
Response: The shock of Madoff led to soul-searching and reform in the
regulatory community. The SEC, embarrassed by missing the scheme,
restructured its examination process, created specialized task forces
(including an Asset Management Unit to scrutinize investment advisors), and
set up a whistleblower office in 2011 to pay rewards for tips – an
acknowledgment that insider information can crack cases that exams miss.
Laws were updated: the Dodd-Frank Act of 2010 expanded SEC powers
and introduced more investor protections. Globally, the Madoff case spurred
other countries to tighten their fund oversight and cooperate more on
cross-border supervision (Madoff had feeder funds and clients worldwide).
On the broader front, the crisis yielded new regulations like the U.S. Fraud Enforcement and Recovery Act (FERA) of 2009 , which bolstered resources for prosecuting financial fraud, and the
creation of the Financial Crisis Inquiry Commission to investigate
wrongdoing that contributed to the meltdown. One clear outcome: a slew of
large banks paid tens of billions in fines and settlements for mortgage and
securities fraud in the following years. The late 2000s taught a harsh
lesson that even in sophisticated markets, basic frauds (Ponzi schemes, false statements) can proliferate if oversight breaks down – and that restoring trust requires
both stronger enforcement and structural reform.
2010s – Cyber Fraud, Hacks, and Crypto Scams
Development: The 2010s saw banking go truly digital – with mobile
apps, instant payments, and cryptocurrencies. This opened new frontiers for
fraud: cyberattacks on banks, online theft of customer data, and
crypto-related scams. The global nature of internet finance meant criminals
from anywhere could target victims everywhere, often anonymously.
Case (Cyber Heist): In February 2016, hackers orchestrated one of the
largest bank robberies in history without stepping foot in a branch.
They broke into Bangladesh Bank’s systems and issued fraudulent instructions
via the SWIFT network to steal $951 million from the bank’s account at the
New York Federal Reserve. While a typo (“fandation” instead of
“foundation”) stopped most of the transfers, $81 million succeeded in reaching accounts in the Philippines
[62]
[63].
The money vanished into the casino industry and hasn’t been fully
recovered. The hack revealed that even central banks could be victims of
cyber fraud, and it exposed weak links in global payment networks. SWIFT
subsequently urged banks to beef up security and launched programs to
certify compliance. Another case: Equifax Hack (2017) – personal
data of 147 million was stolen, fueling identity theft fraud for years. And
as banking went online, account takeovers and phishing attacks on
customers skyrocketed. The FBI reported that in 2021 alone, Americans lost
$2.4 billion to phishing and related scams – the largest category of
cybercrime by victims.
Case (Crypto Fraud): The cryptocurrency boom created a Wild West for
fraudsters. A notable scam was OneCoin (2014–2016), a fake
cryptocurrency run by Ruja Ignatova, the self-styled “Cryptoqueen.” OneCoin
turned out to be a Ponzi scheme without any real blockchain; about 3.5
million victims worldwide invested over $4 billion before it
collapsed
[64]
[65].
Ignatova disappeared in 2017 and remains a fugitive on the FBI’s Most
Wanted list
[66].
Another high-profile case was Bitconnect, a crypto Ponzi that
imploded in 2018 and cost investors $3.5 billion. Additionally, the decade
saw multiple cryptocurrency exchange hacks (Mt. Gox lost ~$450
million in Bitcoin in 2014) and scammers pushing fraudulent Initial Coin
Offerings (ICOs) – by some estimates, over 80% of ICOs in 2017 were scams
or failed projects.
Response: The 2010s forced a paradigm shift: regulators and banks
started treating cyber risks as seriously as traditional fraud. Banks
invested heavily in cybersecurity measures, from multi-factor
authentication for customers to AI-based intrusion detection systems. After
Bangladesh Bank, SWIFT established mandatory security controls for member
banks and created a cyber threat-sharing hub. Regulators issued guidance
and in some cases regulations on cyber hygiene (e.g., NY State’s DFS
cybersecurity rule in 2017 requiring banks to meet specific security
standards). Law enforcement also adapted, creating cyber task forces to
tackle crimes that traverse digital borders. For crypto, by the late 2010s
governments began rolling out crypto-specific rules: e.g.,
anti-money-laundering regulations were extended to crypto exchanges (FinCEN
did so in the US; the EU’s 5th AML Directive did similarly). Several
countries issued investor warnings or bans on certain crypto schemes; and
agencies like the SEC cracked down on fraudulent ICOs, treating many as
unregistered securities. OneCoin’s fallout led Europol and others to
coordinate more on crypto Ponzi investigations, and the fact that Ignatova
was added to the FBI’s Top Ten list in 2022 shows the priority given to
these new-era scammers. By the end of the decade, blockchain analytics firms had emerged, assisting in tracing crypto transactions for law enforcement –
a new toolkit against digital laundering. The overarching lesson: as
finance digitizes, regulations and protective measures must innovate in
tandem. Cyber fraud and crypto scams taught regulators that fraud prevention can no longer be local or analog ; it must be tech-savvy, global, and as decentralized as the threats.
2020s – The Crypto Crash and the AI Frontier
Development: The early 2020s have seen the rise and fall of major
crypto empires and the dawn of widely accessible artificial intelligence.
Decentralized finance (DeFi) and crypto exchanges, some handling tens of
billions in assets, often operated in regulatory grey areas – until
spectacular failures prompted calls for action. Simultaneously, AI tools
(like deepfakes and large language models) grew powerful enough to be
weaponized by scammers, posing new challenges for fraud prevention in
real-time.
Case: In 2022, the collapse of FTX, a leading cryptocurrency
exchange valued at $32 billion, sent shockwaves through the financial world.
FTX’s founder, Sam Bankman-Fried (SBF), was revered in crypto – until it
emerged that customer funds (over $8 billion) were misused to cover
losses at his trading firm, Alameda Research
[67]
[68].
This alleged fraud – essentially using clients’ money as a personal piggy
bank – led to FTX’s bankruptcy in November 2022 and SBF’s arrest. The
scandal vaporized billions in customer assets, affecting over a million
users, and severely dented trust in the crypto industry. It drew parallels
to Enron and Madoff, given the mix of deception and audit failures, and
underscored that crypto institutions can harbor very traditional fraud. In
FTX’s wake, other crypto firms like Celsius and Voyager also collapsed amid
accusations of mismanagement or fraud. Meanwhile, AI-driven fraud stepped up: by 2023, reports showed a surge in deepfake scams, and the FBI
warned that criminals were using deepfake video interviews to penetrate
remote hiring (to then commit insider fraud). One audacious 2023 incident
saw criminals use a deepfake hologram of a company CEO in a video call to
trick employees into transferring funds – a sci-fi level con that was
successful until discovered later.
Response: Regulators worldwide are now sprinting to regulate crypto.
In the U.S., the FTX debacle provoked lawmakers to draft new bills on
crypto exchange oversight, and regulators like the SEC asserted
jurisdiction, charging SBF with securities fraud
[69].
Several countries that were hands-off are enacting rules: the EU approved MiCA (Markets in Crypto-Assets Regulation) in 2023 to bring
exchanges and stablecoins under supervision, and many jurisdictions are
implementing stricter custody and reserve requirements for crypto
platforms to prevent another FTX-type misuse of funds. Bank regulators
issued guidelines limiting banks’ exposures to crypto and requiring risk
management for digital assets. On the AI front, financial authorities have
begun addressing AI in fraud and security contexts. For example, FINRA and
the SEC in the US have highlighted the need for firms to guard against
deepfake-related scams and ensure any AI used in finance is fair and
well-governed. Banks themselves are deploying defensive AI: some now use
deepfake detection for video banking and voice biometrics for call
verification, as mentioned earlier. Government agencies have also ramped up
public advisories – the FTC in 2023 released consumer alerts on
“impersonation scams using AI voice clones.” There is even discussion of
requiring watermarking of AI-generated media to help identify fakes. In
essence, the responses of the 2020s are still evolving, but a clear
trajectory is visible: extending the regulatory perimeter to
encompass crypto finance and adopting new tech-centric approaches to combat
AI-enabled fraud. The lessons are fresh: huge frauds can grow in lightly
regulated spaces (be it crypto or AI-driven manipulation), so authorities
are expanding laws and tools to shine light into those corners.
2025 – What about today and tomorrow?
As we move further into the 2020s, banks and regulators alike will need to
continuously adapt – the fraudsters certainly will. For instance, ChatGPT's new image generator has demonstrated the capability to create fake receipts with convincing
text and formatting. This advancement could potentially be exploited by
fraudsters to produce counterfeit documents for deceptive purposes.
[70]
This risk extends far beyond receipts. As online onboarding becomes the
norm in financial services, especially in mobile banking and fintech apps,
users are often required to upload images of identity documents — such as passports, driver's licenses, utility bills, or diplomas — to verify
who they are. With advanced generative AI tools now capable of producing
photorealistic forgeries, bad actors can fabricate entire digital
identities or impersonate real people with frightening accuracy.
For instance, a fraudster can now generate a fake passport image complete with realistic fonts, holograms, and
watermarks — elements that used to require specialized knowledge or hardware — simply
by prompting an AI tool
[71].
Combine that with deepfake selfies or AI-generated video clips mimicking
someone’s face and voice, and it becomes increasingly difficult for
traditional KYC (Know Your Customer) processes to distinguish between a real
person and a convincing fraud.
This is especially dangerous in remote onboarding scenarios, where there’s
no in-person verification and everything relies on uploaded images and
facial recognition. Banks and fintechs are already reporting an uptick in
synthetic identity fraud, where criminals blend real and fake information
to create seemingly legitimate profiles. Diplomas, employment letters, pay
stubs, and bank statements can also be generated or altered, enabling
fraudsters to pass background checks for loans, credit, or even job offers
in finance — opening the door to insider threats.
Financial regulators are increasingly warning that identity-proofing must
evolve beyond static document checks. There’s growing emphasis on
multi-factor, behavior-based authentication, AI detection of image
manipulation, and collaborations with document-issuing authorities to
cross-check authenticity in real-time. But the threat surface is expanding
fast — and without robust controls, the era of "upload your ID and you're
in" could become an open invitation to sophisticated fraud.
Response: As artificial intelligence (AI) technologies advance,
financial institutions are increasingly adopting sophisticated measures to
combat AI-generated document fraud in digital onboarding processes. These
measures range from widely implemented industry standards to emerging
technologies offering additional security layers.
Standard Measures in Use
Liveness Detection and Biometric Verification: To ensure that a live
person is present during onboarding, institutions employ liveness detection
techniques
[72].
Users may be prompted to perform specific actions, such as blinking or head
movements, during selfie verification. Companies like Onfido integrate
biometric verification technologies into banking apps, guiding users to
capture selfies or videos for identity confirmation.
AI-Powered Document Forensics: Advanced AI models are utilized to
scrutinize identity documents for signs of tampering, such as
inconsistencies in fonts, layouts, or embedded security features. Socure
employs computer vision models to detect alterations in document
submissions, effectively identifying and preventing the use of fake IDs and
stolen identities
[73].
Cross-Referencing with External Databases: Financial institutions
cross-verify user-submitted information with authoritative external
databases, including government records and credit bureaus, to confirm
identity authenticity and detect discrepancies indicative of synthetic
identities.
Participation in Fraud Intelligence Networks: Organizations
collaborate through networks like the Financial Services Information
Sharing and Analysis Center (FS-ISAC) to share real-time information on
emerging threats, enhancing collective defense mechanisms against cyber
threats
[74].
Regulatory Guidance on AI and Fraud Prevention: Regulatory bodies,
such as the Financial Industry Regulatory Authority (FINRA) provide
guidance on addressing AI-driven fraud, including the use of deepfake
technologies. FINRA highlights the potential for scammers to create
realistic fake content, such as deepfake videos, to manipulate stock prices
or promote fraudulent schemes
[75].
Emerging Additional Security Measures
Blockchain-Based Verifiable Credentials: This approach involves
issuing tamper-proof digital credentials stored on a blockchain, allowing
instant verification of documents like diplomas or licenses. Microsoft
Entra Verified ID utilizes decentralized identifiers (DIDs) to
cryptographically sign credentials, ensuring their authenticity
[76].
Cryptographically Signed Identity Documents: Countries like Estonia
have implemented digital ID cards that enable secure authentication and
digital signing for various e-services
[77].
These ID cards are widely accepted by public institutions and preferred by
private sector companies for conducting business electronically.
Content Provenance and Watermarking Standards: The Coalition for
Content Provenance and Authenticity (C2PA) has developed standards to trace
the origin and history of digital assets, embedding provenance data to help
verify content authenticity and detect AI-generated forgeries
[78].
Behavioral Biometrics for Fraud Detection: Technologies like BioCatch
analyze user behavior patterns, such as typing dynamics and mouse
movements, to detect anomalies that may indicate fraudulent activity. This
real-time analysis helps in identifying transactions conducted under the
influence of cybercriminals
[79].
By integrating these standard and emerging measures, financial institutions
can enhance their defenses against the sophisticated threats posed by
AI-generated document fraud.
References
Rogers, J. (2019). Deepfake Voice Fraud: Scammers Use AI to Mimic CEO’s Voice and Steal
$243,000 . Avast. Available: https://blog.avast.com/deepfake-ceo-scam
Internet Crime Complaint Center (IC3). (2019). Business Email Compromise: The $26 Billion Scam . FBI IC3 Public Service Announcement. Available: https://www.ic3.gov/Media/Y2019/PSA190910
Reuters. (2014). Decades-long ties to Madoff cost JPMorgan $2.6 billion . Reuters, 07-Jan-2014. Available: https://www.reuters.com/article/us-jpmorgan-madoff-settlement/decades-long-ties-to-madoff-cost-jpmorgan-2-6-billion-idUSBREA060JL20140108
ACI Worldwide. (2022). Four reasons banks and financial institutions should fight fraud in
real time . ACI Blog. Available: https://www.aciworldwide.com/blog/four-reasons-banks-and-financial-institutions-should-fight-fraud-in-real-time
Galileo Financial Technologies. (2024). How Banks and Fintechs Can Beat Fraudsters in the AI Arms Race . Galileo Blog, 05-Apr-2024. Available: https://www.galileo-ft.com/blog/how-banks-and-fintechs-can-beat-fraudsters-in-the-ai-arms-race/
Investopedia. (2025). America’s First Financial Fraudsters.
Available: https://www.investopedia.com/articles/financial-theory/09/history-of-fraud.asp
Simons, T. (2023). Synthetic identity – a new path for government fraud? . Thomson Reuters Insights, 15-Mar-2023. Available: https://legal.thomsonreuters.com/en/insights/articles/synthetic-identity-fraud
[Kimery, A. (2024). Report: Synthetic identity fraud is growing.
Biometric Update, 15-Oct-2024. Available: https://www.biometricupdate.com/202410/report-synthetic-identity-fraud-is-growing
U.S. Government Accountability Office (GAO). (2022). Estimated Amount of Fraud During Pandemic Likely Between $100 Billion
and $135 Billion . GAO-23-106701, 12-Dec-2022. Available: https://www.gao.gov/products/gao-23-106701
FBI. (2020). Money Mules. FBI Common Frauds & Scams. Available: https://www.fbi.gov/how-we-can-help-you/scams-and-safety/common-frauds-and-scams/money-mules
U.S. Department of the Treasury. (2022). U.S. Treasury Sanctions Notorious Virtual Currency Mixer Tornado Cash . Press Release, 08-Aug-2022. Available: https://home.treasury.gov/news/press-releases/jy0916
The Washington Post. (1991). BCCI Scandal: Behind the 'Bank of Crooks and Criminals' . The Washington Post, 28-Jul-1991. Available: https://www.washingtonpost.com/archive/politics/1991/07/28/bcci-scandal-behind-the-bank-of-crooks-and-criminals/563f2216-1180-4094-a13d-fd4955d59435/
U.S. Secret Service. (2021). Worthy of Trust and Confidence – 150+ Years of History . Available: https://www.secretservice.gov/about/history
FinCEN. (n.d.). History of Anti-Money Laundering Laws. U.S.
Financial Crimes Enforcement Network. Available: https://www.fincen.gov/history-anti-money-laundering-laws
FATF. (2019). History of the FATF. Financial Action Task Force.
Available: https://www.fatf-gafi.org/en/historyofthefatf
Monroe, H., Carvajal, A., & Pattillo, C. (2010). Perils of Ponzis . Finance & Development, vol. 47, no. 1, March 2010. International
Monetary Fund. Available: https://www.imf.org/external/pubs/ft/fandd/2010/03/monroe.htm
Historic UK. (n.d.). The South Sea Bubble of 1720. Historic-UK.com.
Available: https://www.historic-uk.com/HistoryUK/HistoryofEngland/South-Sea-Bubble/
Bubble Act 1720. (1720). (Ann. 6 Geo I) “An Act to restrain the extravagant and unwarrantable
Practice of raising Money by voluntary Subscriptions...” (Bubble Act). Available: https://www.lsd.law/browse/index.php/Legislation/UK/1720/6_Geo_1_%28Bubble_Act%29
Reuters. (2018). Danske Bank's 200 billion euro money laundering scandal . Reuters, 19-Nov-2018. Available: https://www.reuters.com/article/danske-bank-moneylaundering-timeline/danske-banks-200-billion-euro-money-laundering-scandal-idUSKCN1NO10D
AML Graph Analysis with FINaplo.AI. Available: https://www.youtube.com/playlist?list=PLR0D4Y_oFo5tIOc_cIMdbPI1AESqOYPhR
U.S. Department of Justice (SDNY). (2023). Co-Founder Of Multibillion-Dollar Cryptocurrency Scheme “OneCoin”
Sentenced To 20 Years In Prison . Press Release, 12-Sep-2023. Available: https://www.justice.gov/usao-sdny/pr/co-founder-multibillion-dollar-cryptocurrency-scheme-onecoin-sentenced-20-years-prison
Investopedia. (2023). What Happened to OneCoin, the $4 Billion Crypto Ponzi Scheme? . Available: https://www.investopedia.com/terms/o/onecoin.asp
Investopedia. (2023). The Collapse of FTX: What Went Wrong With the Crypto Exchange? . Available: https://www.investopedia.com/what-went-wrong-with-ftx-6828447
Convera. (2023). After the fraud: SBF, FTX and the future of crypto regulations . Convera Insights, 07-Feb-2023. Available: https://www.convera.com/learning-center/article/after-fraud-sbf-ftx-and-future-crypto-regulations
FS-ISAC – Fraud Intelligence and Information Sharing. Available: https://www.fsisac.com/
Microsoft Entra Verified ID – Decentralized Identifiers (DIDs). Available: https://learn.microsoft.com/en-us/entra/verified-id/decentralized-identifier-overview
Estonia e-ID – Digital ID Cards. Available: https://e-estonia.com/solutions/estonian-e-identity/id-card/
Disclaimer: Any mention of a company, product, or service in this
document is provided solely for illustrative purposes as an example or
use case. These references should not be construed as endorsements,
promotions, or recommendations of any kind.