Welcome to Your Briefs
There is no excerpt because this is a protected post.
Algorithmic Bias | Towards greater Transparency, Fairness and Representation in automated systems design
Could you be a victim of algorithmic discrimination without even realising it?
Algorithms are increasingly being used in more settings than ever before and they are often relied upon to make snap decisions on future events, based of patterns found in historic data.
Understandably though, whilst the potential for automation and predictive analytics in decision making is vast; A rise in the adoption of algorithms has also brought with it increased risks of over reliance and social harm as algorithms continue to take an ever more prominent role in our lives.
AMZN.O’s recruiting algorithm and the the use of AI algorithms by the police are just two examples that have helped cast a spotlight on the falibility of automated system design. Most importantly, depending on how they are created, algorithmic errors can to not only reflect present and historic social issues, but aggravate them also. These risks and effects, will need to be closely monitored to prevent repeating human errors in an irreversible way.
Join WiBrief In this latest article where adopt a case study approach to deep dive into recent high profile cases of algorithmic use and misuse:
1. What are algorithms,
2. The algorithmic problem – systematising bias in code,
3. What is ‘Algorithmic Bias’ and the different types of bias,
4. AMZN.O’S recruitment Case study
5. The Legal Context
6.The Social Context
7. Outlook: The need for standardising an approach for detecting and measuring algorithmic Bias.
8. Key takeaways
Read the full article here:
If you enjoyed this article, then look out for our upcoming coverage of the recent changes in AI Regulations, covered by AI WAVES
Read the full article here. …
WiredBrief are delighted to bring to you an exciting event hosted by AI Waves, which looks at Nextbanks AI driven FinTech credit scoring solution and its ground-breaking implications for the financial sector. …
Someone has just very generously gifted you a Drone and you are excited to try it out. You and the Drone are both charged and ready to venture out on your first test flight to the smooth backdrop of Sinatra. Teeming with ideas you step outside. Suddenly however, you remember, …
Top Story: The Information Commissioner’s Office (ICO) has recently just published a new data analytics toolkit. The toolkit is expected to help organisational users of ‘Data analytics’ better understand whether their AI and Data Analytics practises, are compliant with relevant data protection regimes. This brief looks at the ICO’s toolkit and …
WiredBrief are delighted to bring to you an exciting panel event hosted by Legal Hackers MCR, which looks into Blockchain and what the hype is all about. Click here to get you tickets! …
Have you ever wondered how Blockchain could transform the world and how companies are using its transformative power to bring about social change?
The consequences of the recent COVID-19 pandemic have allowed Blockchain to quickly rise in prominence over the past year to become one of the most prolific technologies on the market in 2021, but little is still known about it.
Join us at this exciting event where we will be covering:
1. HOW BLOCKCHAIN WORKS
2. WHY BLOCKCHAIN HAS THE CAPACITY TO MASSIVELY DISRUPT INDUSTRY
3. INSIGHT INTO SOME LEADING PROJECTS AND PRODUCTS IN THIS SPACE …
The European Commission has delivered a new risk-assessment framework for trustworthy Artificial Intelligence (AI).
The new Risk-Assessment Framework provides practical guidance for Developers and Programmers who are building AI. A key consideration of the Risk-Assessment, was the innovations impacts on core human rights.
Check out this latest ‘Brief’ which looks at the main aspects of the European Commission’s Risk-Assessment Framework, outlining its key features whilst also bringing in some analysis from recent developments in tech. …
Manchester Legal Hackers and F-LEX take a look at the world of work in life after COVID-19. Have you ever tried to explain your career plans or studies to your parents or grandparents? It’s common to experience blank, confused or amused reactions because each generation has seen huge shifts in …
Are you interested in finding out more about: the applications of Legal Technology to delivering social justice during the Covid 19 pandemic, the ICO’s approach to the regulation of novel IT solutions, such as the UKs recently created contract tracing app, or how the Manchester law and technology initiative is …
The ‘Notre-Dame tipping point’ – A study of key events in Biotechnology, what it means to be alive and whether things will actually ever be the same once the ‘fire’ has passed
Break through developments in Biotechnology concerning the recent restoration of ‘pig brains’ at Yale and the growth of organic robots at Cornell suggest that what it means to be ‘alive’ and human might not be as clearly cut out as first thought.
This ‘Brief’ delves into the studies placing them into their legal and social contexts whilst providing perspective on the difficult questions and issues they also raise.
This latest article takes a look at current UK road traffic law and examines its suitability for regulating vehicle accidents involving self-driving vehicles. But notably whilst the law has traditionally been focused on placing liability on human ‘drivers’ who have until recently been at the heart of most road traffic accidents, would it still be right or even fair to argue the same in light of the growing capabilities of highly autonomous self-driving vehicles? …
An introduction to the five stages of automation, beginning from 0, where the human driver is in control of the vehicle through to stage five where complete control is handed over to the on-board AI. …
Managing Expectations: The Road to Driverless Cars and the Contributory Roles of Human and Machine Error
A Road Map for Driverless Cars provided by engineers has estimated that fully driverless cars could be on our roads from as early as 2030 but should we be worried? …