MatthiasPi commited on
Commit
adbdca7
1 Parent(s): ffd9d26
Files changed (1) hide show
  1. README.md +9 -3
README.md CHANGED
@@ -1,3 +1,11 @@
 
 
 
 
 
 
 
 
1
  # Representativity-based active learning for regression using Wasserstein distance and GroupSort Neural Networks
2
 
3
  You will find in this repository the codes used to test the performance of the WAR model on a fully labeled dataset
@@ -19,6 +27,4 @@ You will find in this repository the codes used to test the performance of the W
19
 
20
 
21
  ## Abstract
22
- This paper proposes a new active learning strategy called Wasserstein active regression (WAR) based on the principle of distribution-matching to measure the representativeness of our labeled dataset compared to the global data distribution. We use GroupSort Neural Networks to compute the Wasserstein distance and provide theoretical foundations to justify the use of such networks with explicit bounds for their size and depth. We combine this solution with another diversity and uncertainty-based approach to sharpen our query strategy. Finally, we compare our method with other solutions and show empirically that we consistently achieve better estimations with less labeled data.
23
-
24
-
 
1
+ ---
2
+ datasets:
3
+ - maxcembalest/boston_housing
4
+ language:
5
+ - en
6
+ tags:
7
+ - code
8
+ ---
9
  # Representativity-based active learning for regression using Wasserstein distance and GroupSort Neural Networks
10
 
11
  You will find in this repository the codes used to test the performance of the WAR model on a fully labeled dataset
 
27
 
28
 
29
  ## Abstract
30
+ This paper proposes a new active learning strategy called Wasserstein active regression (WAR) based on the principle of distribution-matching to measure the representativeness of our labeled dataset compared to the global data distribution. We use GroupSort Neural Networks to compute the Wasserstein distance and provide theoretical foundations to justify the use of such networks with explicit bounds for their size and depth. We combine this solution with another diversity and uncertainty-based approach to sharpen our query strategy. Finally, we compare our method with other solutions and show empirically that we consistently achieve better estimations with less labeled data.