
{"id":88669,"date":"2025-08-15T09:20:24","date_gmt":"2025-08-15T09:20:24","guid":{"rendered":"https:\/\/mycryptomania.com\/?p=88669"},"modified":"2025-08-15T09:20:24","modified_gmt":"2025-08-15T09:20:24","slug":"eu-ai-act-article-23-understanding-obligations-of-importers-of-high-risk-ai-systems","status":"publish","type":"post","link":"https:\/\/mycryptomania.com\/?p=88669","title":{"rendered":"EU AI Act Article 23: Understanding Obligations of Importers of High-risk AI Systems"},"content":{"rendered":"<p>European Union (EU) Artificial Intelligence (AI) Act (Regulation (EU) 2024\/1689) represents the first most comprehensive regulation on AI. It proposes a proactive framework to regulate AI systems and minimise the risk of AI systems against to the health, safety, or fundamental rights of end users. Article 23 establishes obligations for importers of high-risk AI systems to ensure these systems comply with EU standards before and after entering the market. At its core, the article conceptualizes importers as key gatekeepers in the AI supply chain, bridging non-EU providers (manufacturers or developers) and the EU market. Importers are not mere intermediaries but active participants accountable for verifying conformity, mitigating risks, and facilitating regulatory oversight. This aligns with the Act\u2019s broader goal of promoting safe, trustworthy AI while protecting fundamental rights, health, and\u00a0safety.<\/p>\n<p>The \u201cimporter\u201d is defined in the Act as a natural or legal person located or established in the Union (EU) that places on the market an AI system bearing the name or trademark of a natural or legal person established in a third country (outside the EU). This definition emphasizes the importer\u2019s role in introducing foreign-made AI systems into the EU, making them liable for compliance even if the original provider is extraterritorial.<\/p>\n<p>The article\u2019s obligations can be conceptualized into three phases: pre-market verification (ensuring readiness before market entry), ongoing responsibility (during storage, transport, and handling), and post-market cooperation (record-keeping and interaction with authorities). These elements create a layered accountability system, where importers must proactively check, react to issues, and support enforcement. Below, I\u2019ll explain each element clearly, drawing from the article\u2019s seven paragraphs.<\/p>\n<p>Grok<strong>Pre-Market Verification (Article\u00a023(1))<\/strong><\/p>\n<p>Core Obligation: Before placing a high-risk AI system on the market, importers must verify its conformity with the EU AI\u00a0Act.<\/p>\n<p>Key Elements to\u00a0Check:<\/p>\n<p>The provider has completed the relevant conformity assessment procedure (as outlined in Article 43, which involves third-party audits for high-risk systems).The provider has prepared technical documentation (per Article 11 and Annex IV, detailing design, development, and risk management).The system has the CE marking (indicating EU compliance) and is accompanied by the EU declaration of conformity (Article 47) and user instructions.The provider has appointed an authorized representative in the EU (Article 22(1)) to handle regulatory matters.<\/p>\n<p>Explanation: This acts as a \u201cdue diligence checkpoint\u201d to prevent non-compliant systems from entering the EU. Importers aren\u2019t required to perform assessments themselves but must confirm the provider\u2019s work, reducing the risk of unsafe AI (e.g., biased facial recognition tools) reaching\u00a0users.<\/p>\n<p><strong>2. Handling Non-Conformity (Article\u00a023(2))<\/strong><\/p>\n<p>Core Obligation: If there\u2019s reason to believe the system is non-compliant, falsified, or has fake documentation, the importer must not place it on the market until it\u2019s\u00a0fixed.<\/p>\n<p>Key Elements:<\/p>\n<p>Assess risks: If the system poses a risk (as defined in Article 79(1), e.g., threats to health or rights), inform the provider, authorized representative, and market surveillance authorities.Proactive remediation: Bring the system into conformity before proceeding.<\/p>\n<p>Explanation: This element introduces a \u201crisk escalation\u201d mechanism, empowering importers to halt distribution and notify stakeholders. It conceptualizes importers as risk managers, ensuring accountability flows back to the provider while protecting the EU market from immediate harms.<\/p>\n<p><strong>3. Identification and Traceability (Article\u00a023(3))<\/strong><\/p>\n<p>Core Obligation: Importers must add their contact details (name, trade name\/mark, address) to the AI system, its packaging, or documentation.<\/p>\n<p>Key Elements:<\/p>\n<p>Applies where applicable (e.g., physical products or digital interfaces).Ensures traceability for users, regulators, and supply chain\u00a0actors.<\/p>\n<p>Explanation: This fosters transparency and accountability, making it easier to contact the importer for complaints, recalls, or investigations. It\u2019s a simple yet crucial element in the Act\u2019s emphasis on identifiable responsibility in the AI value\u00a0chain.<\/p>\n<p><strong>4. Storage and Transport Conditions (Article\u00a023(4))<\/strong><\/p>\n<p>Core Obligation: While the system is under the importer\u2019s control, ensure storage or transport doesn\u2019t compromise compliance with the Act\u2019s requirements (Section 2, covering aspects like robustness, accuracy, and cybersecurity).<\/p>\n<p>Key Elements:<\/p>\n<p>Monitor conditions (e.g., temperature, handling) that could affect the system\u2019s integrity.Applies only where relevant (e.g., hardware-based AI).<\/p>\n<p>Explanation: This conceptualizes the importer\u2019s role extending beyond paperwork to physical stewardship, preventing degradation that could turn a compliant system non-compliant (e.g., data corruption in transit affecting AI performance).<\/p>\n<p><strong>5. Record-Keeping (Article\u00a023(5))<\/strong><\/p>\n<p>Core Obligation: Retain key documents for 10 years after the system is placed on the market or put into\u00a0service.<\/p>\n<p>Key Elements:<\/p>\n<p>Copies of: Notified body certificate (if applicable), instructions for use, and EU declaration of conformity (Article\u00a047).Retention period aligns with potential long-term risks in AI\u00a0systems.<\/p>\n<p>Explanation: This element supports auditability and long-term enforcement, allowing regulators to retroactively verify compliance. It underscores the Act\u2019s forward-looking approach to AI, where issues might emerge years later (e.g., evolving biases in machine learning\u00a0models).<\/p>\n<p><strong>6. Information Provision to Authorities (Article\u00a023(6))<\/strong><\/p>\n<p>Core Obligation: Upon reasoned request, provide competent authorities with all necessary information and documentation to prove conformity.<\/p>\n<p>Key Elements:<\/p>\n<p>Include records from Paragraph 5 and ensure technical documentation is available.Deliver in an easily understandable language.<\/p>\n<p>Explanation: This facilitates regulatory inspections, conceptualizing importers as cooperative partners in oversight. It ensures authorities can access evidence without barriers, enhancing the Act\u2019s enforcement teeth.<\/p>\n<p><strong>7. Cooperation with Authorities (Article\u00a023(7))<\/strong><\/p>\n<p>Core Obligation: Assist competent authorities in actions related to the imported system, especially to reduce or mitigate\u00a0risks.<\/p>\n<p>Key Elements:<\/p>\n<p>Broad cooperation: Includes investigations, corrective measures, or\u00a0recalls.Focus on risk mitigation for high-risk AI.<\/p>\n<p>Explanation: As the final element, this ties everything together by mandating active support, turning importers into allies in public protection. It reflects the Act\u2019s collaborative ecosystem, where private actors help enforce public interest\u00a0goals.<\/p>\n<p><a href=\"https:\/\/medium.com\/coinmonks\/eu-ai-act-article-23-understanding-obligations-of-importers-of-high-risk-ai-systems-6d427937809c\">EU AI Act Article 23: Understanding Obligations of Importers of High-risk AI Systems<\/a> was originally published in <a href=\"https:\/\/medium.com\/coinmonks\">Coinmonks<\/a> on Medium, where people are continuing the conversation by highlighting and responding to this story.<\/p>","protected":false},"excerpt":{"rendered":"<p>European Union (EU) Artificial Intelligence (AI) Act (Regulation (EU) 2024\/1689) represents the first most comprehensive regulation on AI. It proposes a proactive framework to regulate AI systems and minimise the risk of AI systems against to the health, safety, or fundamental rights of end users. Article 23 establishes obligations for importers of high-risk AI systems [&hellip;]<\/p>\n","protected":false},"author":0,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[2],"tags":[],"class_list":["post-88669","post","type-post","status-publish","format-standard","hentry","category-interesting"],"_links":{"self":[{"href":"https:\/\/mycryptomania.com\/index.php?rest_route=\/wp\/v2\/posts\/88669"}],"collection":[{"href":"https:\/\/mycryptomania.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/mycryptomania.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/mycryptomania.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=88669"}],"version-history":[{"count":0,"href":"https:\/\/mycryptomania.com\/index.php?rest_route=\/wp\/v2\/posts\/88669\/revisions"}],"wp:attachment":[{"href":"https:\/\/mycryptomania.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=88669"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/mycryptomania.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=88669"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/mycryptomania.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=88669"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}