Please note: The algorithm descriptions in English have been automatically translated. Errors may have been introduced in this process. For the original descriptions, go to the Dutch version of the Algorithm Register.
BIG re-registration selection model
- Publication category
- Impactful algorithms
- Impact assessment
- DPIA
- Status
- In use
General information
Theme
- Work
- Health and Healthcare
Begin date
Contact information
Link to publication website
Responsible use
Goal and impact
Considerations
Assigning a risk category by the selection model aims to reduce the administrative burden on healthcare providers and work more on the basis of trust rather than mistrust. Reducing the administrative burden on healthcare providers requires assigning them to risk categories. Reducing the administrative burden is difficult to achieve any other way.
Human intervention
The selection model determines whether an application is selected for having to provide (objectively verifiable) documentary evidence.
The use of the algorithm does not lead to fully automated decision-making. After all, the final substantive assessment of an application based on the evidence provided is done by an employee.Risk management
- Preventing direct discrimination - when the files are assessed by the algorithm, no sensitive file characteristics (such as gender, age or having a foreign nationality) are shared on which the algorithm could make a direct distinction. Because these are not included as file characteristics, the algorithm cannot select on them either. Moreover, these characteristics are also not included as relevant criteria in the decision trees. This is to prevent direct discrimination on these characteristics;
- Preventing indirect discrimination - although the decision tree cannot select directly on sensitive file characteristics, it is theoretically possible that a certain - seemingly innocent - characteristic is included in the tree, which has the practical effect that certain groups of healthcare providers are more likely to receive a high risk score. This is prohibited when there is no justification for doing so. During the development of the algorithm, this was considered and it was decided to show 'bias graphs' on the dashboard for the generated decision trees and the current decision tree, showing whether, as a result of a decision tree, certain groups of healthcare providers are selected relatively more often than other groups of healthcare providers. This is made transparent for the following characteristics: age, gender and having or not having a foreign nationality. Based on these bias graphs, a tree with the lowest possible bias or a tree where the indirect bias is objectively justifiable can be chosen;
- In addition, the algorithm should be prevented from becoming self-validating. Therefore, in addition to the purposive sample, a random sample also takes place. The subsequent decisions are used to generate new decision trees;
- It is not clear to the staff assessing the supporting documents whether the relevant application fell into the random or targeted selection. This is to prevent them from being unknowingly influenced by this information;
- Regarding explainability, the following is noted: the decision trees clearly show the different threshold values. These threshold values are clear, concrete and also explainable in application to a concrete case;
- Finally, it is noted that a DPIA has been prepared; it has been submitted to the FG for advice and subsequently approved.
Legal basis
The explanatory memorandum includes the following in this regard: 'Applications for periodic registration will not have to be substantiated with supporting documents as standard. The self-declarations will be checked for content based on a random or targeted sample.
Links to legal bases
- Wet BIG: https://wetten.overheid.nl/BWBR0006251
- Besluit periodieke registratie Wet BIG: https://wetten.overheid.nl/BWBR0024841
Impact assessment
Operations
Data
- General information about the case/application
- Personal information (this information is needed to understand any indirect bias)
- Information about hours worked
- Information regarding the content of the work performed
- Measures applicable abroad
Technical design
The algorithm used in targeted selection is a type of "decision tree classifier". It can be seen as a structured method of indicating which applications are likely to be at higher risk of not meeting the legal requirements for re-registration. Each node in this tree represents a specific question about the application, for example the total number of hours worked in the past five years. Based on the answers to these questions, the tree guides the decision across several branches until finally a conclusion is reached: either "select" or "do not select". In training, the algorithm learns to ask the most informative questions at each step based on the available data in order to divide the applications into smaller and more homogeneous groups.
The algorithm offers a choice of 16 different trees. Four different levels in terms of tree depth can be selected (3, 4, 5 or 6) and then the minimum size of the group is determined (100, 200, 300 or 400). Depth refers to the maximum number of splitting or decision steps the tree can have between the root (the start point) and the leaves (the end point). The minimum size of a group refers to the minimum number of applications for BIG re-registration that must be present on an end leaf of the decision tree.
The algorithm calculates the probability that an application belongs to a certain category, allowing us to classify applications as "medium risk" or "high risk". Only "high risk" applications are selected for verification in the targeted sample. This choice was made to ease the administrative burden on healthcare providers as much as possible ('low or medium risk' applications can still fall into the random sample).External provider
Similar algorithm descriptions
- The AP uses this algorithm to classify data breach reports by severity. Based on that classification, inspectors can prioritise serious reports. The algorithm does not contain any personal data.Last change on 11th of October 2024, at 9:33 (CET) | Publication Standard 1.0
- Publication category
- Other algorithms
- Impact assessment
- Field not filled in.
- Status
- In use
- The reporting system's algorithm recognises words in reports, such as 'rubbish' or 'pavement', and automatically determines the correct category and department. As a result, reporters no longer have to choose a category, and reports are dealt with faster at the right department.Last change on 7th of January 2025, at 13:02 (CET) | Publication Standard 1.0
- Publication category
- Other algorithms
- Impact assessment
- DPIA
- Status
- In use
- Investigability Algorithm 'Smart check'. A tool that helps staff determine whether a life support application is research-worthy.Last change on 4th of December 2024, at 16:28 (CET) | Publication Standard 1.0
- Publication category
- Impactful algorithms
- Impact assessment
- DPIA, IAMA
- Status
- In development
- Selection system for selecting contractors to be invited for multiple negotiated procurement of infrastructure works.Last change on 15th of July 2024, at 13:38 (CET) | Publication Standard 1.0
- Publication category
- Impactful algorithms
- Impact assessment
- Field not filled in.
- Status
- In use
- Algorithm that can help registrants avoid "look-alike fraud".Last change on 13th of January 2025, at 8:11 (CET) | Publication Standard 1.0
- Publication category
- Other algorithms
- Impact assessment
- Field not filled in.
- Status
- In use