The Diverging Paths of AI Regulation
The Google Pixel 6 launched recently. Among its many most anticipated features are new artificial intelligence and machine learning features (“AI”), including its speech translation and recognition, and its AI photo editing software. Indeed, although the phone has significant hardware advancements, most commentators recognized that AI advancements were the driving factor behind its success.
This focus on the benefits of AI follows a longer societal trend where there is increasing recognition that AI has countless untapped benefits. Whether it was Alphazero demonstrating new playing styles in chess, dramatically improving efficiency in insurance writing and claims processing, or countries using facial recognition to monitor their citizens, AI is and will continue to fundamentally change the world.
However, countries’ recognition and reaction to AI has not been consistent. Europe, as expected because of its strong stance on individual data rights and privacy, has launched the most aggressive and restrictive stance against the use of AI. Back in 2018, the European Economic Area passed the General Data Protection Regulation, restricting the use of some automated decision making. Additionally, in April 21, 2021, the European Union released a draft of the EU Artificial Intelligence Act, which further attempted to regulate AI through harmonized rules within the European Union. The breadth of the proposed rules is broad – it applies to those outside of the EU that market or provide AI systems to the EU – and the definition of AI is broadly defined, encompassing processes which could reasonably be considered AI.
The proposed rules separate AI into three different tiers of risk: unacceptable risk AI systems, high risk systems, and limited and minimal risk AI systems. Unacceptable risk AI systems such as social scoring or real time remote biometric identification systems are fully banned under the proposed laws. High risk AI systems, including systems that evaluate a consumer’s creditworthiness or use biometric identification in non-public spaces require company oversight, including conducting audits that are similar to Data Protection Impact Assessments to ensure that the systems perform as intended and are secured. Low risk systems continue to have little oversight as the authorities are less concerned about potential abuses with this AI.
On the other end of the spectrum, China has fully embraced the use of AI. Instead of worrying about any negative privacy implications, China has leaned on AI as a tool to build its society and government. Among the uses getting most coverage are China’s use of facial recognition and other AI methods to keep tabs on its citizens, such as the use of AI emotion-detection software on Uyghurs. More generally however, China has woven AI into its social fabric by using it for everyday operations including its social credit system which monitors its citizens and rewards them or punishes them for things they do, its payment and communications systems, or even its defense systems. This general acceptance for AI has been backed by formal legislation such as the ”Made in China 2025” plan or “Next Generation Artificial Intelligence Development Plan”. The effect has been a boom in the research, use, and acceptance of AI (e.g., where as China previously lagged behind in AI research, China has now become the frontrunner.)
Meanwhile, the United States stands in the middle of these two extremes. Like the EU, the United States has agreed that AI should be used in ways that are “based on shared democratic values, including respect for human rights”. Significantly, the U.S. and EU agreed that AI should not be used for social credit scoring. However, the U.S. does not seem to share the EU’s concern over the other potentially invasive and threatening ways that AI could be used and has not publicly committed to a robust federal framework that addresses these AI issues. Instead, the U.S. appears to be concerned over the strategic and geopolitical issues that advanced AI will present, especially if other actors like China become world leaders.
Thus, because of the significant developments in AI and what those developments mean, all countries have been forced reckon with AI regulation. However, geopolitical, historical, and other regulatory forces have created responses that dramatically differ throughout the world. These responses have not only changed the trajectory of AI development in various parts of the world but also increasingly left out the possibility of harmonious AI regulation.
For additional information, please contact us.
FinCEN and Real Estate: Additional Disclosure Requirements May Be On the Horizon for Real Estate Transactions￼
As part of the anti-money laundering regime under the Bank Secrecy Act of 1970 (the “BSA”), in late 2021, the Financial Crimes Enforcement Network (“FinCEN”) division of the Department of the Treasury issued an advanced notice of proposed rulemaking (“ANPRM”) seeking to address potential money laundering through real estate transactions. The comment period for the ANPRM closed on February 21, 2022. This ANPRM comes closely after the notice of proposed rulemaking related to the implementation of the Corporate Transparency Act (the “CTA”), which you can read more about here. Both the CTA and the proposed regulations under the ANPRM would require significant levels of disclosure regarding the beneficial ownership of companies and real estate in non-financed real estate transactions. These measures aim to reduce money laundering, and assets held by undisclosed foreign investors. It is estimated that between 2015 and 2020, at least $2.3 billion was laundered through U.S. real estate, though the actual figure is likely much higher Accordingly, both FinCEN and Congress are trying to limit the number of real estate transactions used to launder money.
Two new laws are set to take effect in the coming months that will require employees to examine their current practices and make changes to bring themselves into compliance.
When I joined Milgrom & Daskam at the height of COVID, I wasn’t sure what the future would look like for me or this relatively young firm. We were giving up our physical office space in downtown Denver and embarking on a new vision for remote workers. Up until then, much of my professional work life was spent in an office environment, surrounded by colleagues My days were punctuate by in-person meetings–formal, over coffee or meals.in the hallways–and bookended by my daily commute between Denver and Los Angeles which ranged anywhere from just under 30 minutes to more than an hour.