Cryptocurrency Derivatives Series: Brazil’s SEC CVM Orders Cryptocurrency-Exchange Binance To Stop Futures Trading
FxWirePro: The Bullish/Bearish Driving Forces of EUR/JPY, OTC Indications & Options Hedging Strategy
Retirement age is increasing – but our new study reveals most only work ten years in good health after 50
FxWirePro: Bullion OTC Bids Indicate Further Upside Risks And Uphold Long Hedges Via Options Strategy
Digital Currency Revolution Series: US CFTC Recognizes Digital Transformation And Hints Holistic Framework On Digital Assets
Digital Currency Revolution Series: Travala.com Partners With Expedia Group For Cryptocurrency-Based Travel Booking
FxWirePro: Risk-Aversion Sentiment To Renew Kiwis Dollar’s Bearishness – Bid NZD/USD 3M IV Skews And Deploy Hedging Strategies
FxWirePro: The Bullish/Bearish Driving Forces of AUD/USD, OTC Indications & Options Strategy For H2’2020
Does your AI discriminate?
But men are chosen as leaders of government around the world in vastly greater numbers.
This disparity is not confined to political leadership. In 2019, Forbes choose 100 of America’s “Most Influential Leaders,” and 99 of them were men.
The lack of diversity is not limited to gender. A survey of nonprofit sector chief executives found that 87% of survey respondents self-identified as white.
As the executive and academic director of a leadership center, I study employment discrimination and inclusion. I’ve seen that many organizations want a process where bias could be removed from identifying leaders. Investors want to invest in businesses with diverse workforces, and employees want to work in diverse organizations.
My research indicates that relying on data analytics to eliminate human bias in choosing leaders won’t help.
AI isn’t foolproof
Employers increasingly rely on algorithms to determine who advances through application portals to an interview.
As labor rights scholar Ifeoma Ajunwa writes, “Algorithmic decision-making is the civil rights issue of the 21st century.” In February 2020, the U.S. House of Representatives’ Committee on Education and Labor convened a hearing called “The Future of Work: Protecting Workers’ Civil Rights in the Digital Age.”
Hiring algorithms create a selection process that offers no transparency and is not monitored. Applicants struck from an application process – or as Ajunwa refers to it, “algorithmically blackballed” – have few legal protections.
For instance, in 2014, Amazon reportedly began developing a computer-based program to identify the best resumes submitted for jobs. The idea was to automate a process and gain efficiency, much as it has done with other aspects of its business.
However, by using computer models to observe patterns in the previous 10 years of submitted resumes to choose the best, the computer taught itself that resumes from men were preferred to a resume that included the word “women’s,” as in a women’s club or organization. Amazon subsequently abandoned the project, according to reports.
Although often historic biases are inadvertently built into algorithms and reflect human prejudices, recent scholarship by Philip M. Nichols has identified an additional threat of potential intentional manipulation of underlying algorithms to benefit third parties.
Inadvertent or intentional, the ability to detect bias of an algorithm is extremely difficult because it can occur at any stage of the development of AI, from data collection to modeling.
Therefore, although organizations have access to leadership analytical tools based on research and analysis of leadership traits, the white male leader stereotype is deeply ingrained and even sometimes perpetuated by those who themselves are diverse. This cannot be eliminated simply by developing an algorithm that selects leaders.
After the interviews
The data to build these algorithms increase exponentially.
One video interview service, HireVue, boasts of its ability to detect thousands of data points in a single 30-minute interview, from sentence structure to facial movements, to determine employability against other applicants.
Imagine the opportunity, then, for a current employer to collect data continuously to determine leadership potential and promotions of its workforce. For instance, cameras in the workplace can collect facial expressions all day at work, particularly when entering and exiting the workplace.
Increasingly, the data are not just collected during the work day or while at work, but during off-duty conduct as well. In a recent article, Professor Inara Scott identified workplace programs that gathered massive amounts of data of off-duty conduct of employees from Facebook posts and Fitbit usage, for example, without transparency about future use of the data. Employers then used those bits of data to draw correlations to predict workplace success.
As Scott notes, most workers “will likely chafe at the notion that their taste in beer, love of indie rock and preference for the Washington Post, along with thousands of other variables, can be used to determine professional development opportunities, leadership potential and future career success.”
Nonetheless, that potential exists today in workplaces, and the law simply has not caught up to the vast amount of data collected and utilized by employers wanting to know the promotion and leadership investment in its employees is supported by the data.
In many cases, employees agree to collection of meta-data without a thorough understanding of what that data can reveal and how it can be used to help or hamper a career.