AI’s Role in Australian Hiring: Unveiling the Legal Gaps

Australia Urged to Lift Veil on AI Use in Hiring as Legal Grey Zone Widens

Australian employers are increasingly turning to artificial intelligence in their hiring processes. However, a significant legal gap exists, as there is currently no requirement for employers to inform candidates when algorithms—not humans—are making pivotal decisions regarding their careers.

The Need for Regulation

This situation has been described as strange and increasingly unsustainable, particularly in light of the rapid adoption of AI in recruitment. Research from the University of Melbourne indicates that approximately 62% of Australian organizations utilize AI in their hiring processes, a figure that climbs even higher among global businesses. Despite this, there is no explicit law mandating disclosure of AI use in job advertisements or hiring practices.

Experts argue that it is essential to introduce some form of regulation regarding AI in recruitment. This is crucial not only because of the significance of hiring processes but also due to individuals’ basic right to earn a living. It raises concerns about how employers using AI account for biases present in algorithms.

Current Legal Frameworks

While there is no dedicated AI disclosure regime for private employers, some existing laws tangentially relate to AI in recruitment.

For example, privacy reforms are being considered that could intersect with hiring practices, especially concerning how personal information is collected, processed, and disclosed. However, these reforms may not fully address the detrimental effects caused by relying solely on AI.

Government agencies are further along in this regard, as they are required to issue transparency statements when using high-risk automated systems, including hiring tools. These statements must disclose the effects on individuals, associated risks, and safeguards in place. Unfortunately, these obligations do not extend to the private sector.

Work health and safety laws provide an indirect means of accountability, requiring employers to manage AI risks similarly to other workplace hazards. However, this has yet to be adequately tested in the context of hiring practices.

Rising Disputes on the Horizon

As the legal landscape stands, it is anticipated that the current vacuum will not last long. Discussions are already underway regarding the necessity for laws to address algorithmic bias in hiring.

For example, the University of Melbourne has cautioned that AI could exacerbate issues of discrimination in hiring. This concern is at odds with Australia’s substantial reliance on such technology. As a result, legal disputes regarding unfair treatment in recruitment may soon arise.

Under the Fair Work Act, applicants can challenge unfair treatment if they believe they have been discriminated against during the hiring process. Some cases in the U.S. have already involved algorithmic decision-making, providing a potential precedent for future Australian cases.

Future Legislative Directions

There remains uncertainty about whether AI in recruitment will come under a broad, EU-style AI Act or be addressed through existing laws. Experts suggest that rather than a sweeping new statute, AI in hiring could be incorporated as a component of the Privacy Act or through an addendum to the Fair Work Act.

Benefits Versus Risks

Despite the risks associated with AI in recruitment, it is recognized as significantly valuable when utilized effectively. AI can help employers quickly process large volumes of applications, identify baseline qualifications, and streamline initial screenings.

However, the same features that enhance efficiency can also exacerbate inequities if left unregulated. AI tools may inadvertently disadvantage groups such as women, individuals whose first language is not English, younger and older workers, and those with disabilities. The potential for filtering out candidates based on historical data that reflects past biases is a significant concern.

Moreover, AI could favor applicants who utilize AI tools to refine their resumes and cover letters, which raises questions about the authenticity of their qualifications.

The Importance of Human Judgment

While automation presents promising opportunities, it is crucial that human judgment remains a vital component of the hiring process. AI should not replace the insights gained from interviews and interpersonal interactions. Perfect candidates on paper may not always fit the organizational culture or team dynamics.

As employers increasingly integrate AI into job advertisements and screening processes, the key message is twofold: embrace the efficiency and insights AI provides, but ensure transparency in its usage, actively manage biases, and keep human decision-making at the forefront.

Without these measures, Australia risks allowing silent algorithms to dictate careers and livelihoods within a growing legal grey zone, with intensified regulation and litigation on the horizon.

More Insights

Revolutionizing Drone Regulations: The EU AI Act Explained

The EU AI Act represents a significant regulatory framework that aims to address the challenges posed by artificial intelligence technologies in various sectors, including the burgeoning field of...

Revolutionizing Drone Regulations: The EU AI Act Explained

The EU AI Act represents a significant regulatory framework that aims to address the challenges posed by artificial intelligence technologies in various sectors, including the burgeoning field of...

Embracing Responsible AI to Mitigate Legal Risks

Businesses must prioritize responsible AI as a frontline defense against legal, financial, and reputational risks, particularly in understanding data lineage. Ignoring these responsibilities could...

AI Governance: Addressing the Shadow IT Challenge

AI tools are rapidly transforming workplace operations, but much of their adoption is happening without proper oversight, leading to the rise of shadow AI as a security concern. Organizations need to...

EU Delays AI Act Implementation to 2027 Amid Industry Pressure

The EU plans to delay the enforcement of high-risk duties in the AI Act until late 2027, allowing companies more time to comply with the regulations. However, this move has drawn criticism from rights...

White House Challenges GAIN AI Act Amid Nvidia Export Controversy

The White House is pushing back against the bipartisan GAIN AI Act, which aims to prioritize U.S. companies in acquiring advanced AI chips. This resistance reflects a strategic decision to maintain...

Experts Warn of EU AI Act’s Impact on Medtech Innovation

Experts at the 2025 European Digital Technology and Software conference expressed concerns that the EU AI Act could hinder the launch of new medtech products in the European market. They emphasized...

Ethical AI: Transforming Compliance into Innovation

Enterprises are racing to innovate with artificial intelligence, often without the proper compliance measures in place. By embedding privacy and ethics into the development lifecycle, organizations...

AI Hiring Compliance Risks Uncovered

Artificial intelligence is reshaping recruitment, with the percentage of HR leaders using generative AI increasing from 19% to 61% between 2023 and 2025. However, this efficiency comes with legal...