Updating content
Browse files- templates/about.html +20 -8
- templates/index.html +52 -11
templates/about.html
CHANGED
@@ -181,18 +181,30 @@
|
|
181 |
</div>
|
182 |
<div class="section">
|
183 |
<div class="section-title">Motivation</div>
|
|
|
|
|
|
|
|
|
|
|
184 |
<p>
|
185 |
Benchmarks usually compare models with <b>many questions</b> from <b>a single minimal context</b>, e.g. as multiple choices questions.
|
186 |
-
This kind of evaluation is little informative of LLMs' behavior in deployment
|
187 |
-
We argue that <b>context-dependence</b> can be seen as a <b>property of LLMs</b>: a dimension of LLM comparison alongside others like size, speed, or knowledge.
|
|
|
188 |
We evaluate LLMs by asking the <b> same questions </b> from <b> many different contexts </b>.
|
189 |
</p>
|
190 |
<p>
|
191 |
-
LLMs are often used to simulate personas and populations.
|
192 |
-
We study
|
193 |
-
To do that we leverage the psychological methodology to study the interpersonal stability of personal value expression
|
194 |
-
We adopt the Schwartz Theory of Basic Personal Values
|
195 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
196 |
</p>
|
197 |
</div>
|
198 |
<div class="section">
|
@@ -234,7 +246,7 @@
|
|
234 |
<div class="section-title">Validation</div>
|
235 |
<p>
|
236 |
Validity refers to the extent the questionnaire measures what it purports to measure.
|
237 |
-
It can be seen the questionnaire's accuracy in measuring the intended factors, i.e. values.
|
238 |
Following the recommendations in <a href="https://pubmed.ncbi.nlm.nih.gov/22329443/">this paper</a>,
|
239 |
the validation consists of two phases: Theory-Based Multidimensional Scaling (MDS) and Confirmatory Factor Analysis (CFA).
|
240 |
</p>
|
|
|
181 |
</div>
|
182 |
<div class="section">
|
183 |
<div class="section-title">Motivation</div>
|
184 |
+
<p>
|
185 |
+
LLMs can role-play different personas by simulating their values and behavior, but can they stick to their role whatever the context?
|
186 |
+
Is simulated Joan of Arc more tradition-driven than Elvis?
|
187 |
+
Will it still be the case after playing chess?
|
188 |
+
</p>
|
189 |
<p>
|
190 |
Benchmarks usually compare models with <b>many questions</b> from <b>a single minimal context</b>, e.g. as multiple choices questions.
|
191 |
+
This kind of evaluation is little informative of LLMs' behavior in deployment, where they are exposed to new contexts (especially considering the LLMs highly context-dependant nature).
|
192 |
+
We argue that <b>undesired context-dependence</b> can be seen as a <b>property of LLMs</b>: a dimension of LLM comparison alongside others like size, speed, or knowledge.
|
193 |
+
While some context-dependence is desired (e.g. following instructions), some is undesired (e.g. drastically changing the simulated value expression based on the interlocutor).
|
194 |
We evaluate LLMs by asking the <b> same questions </b> from <b> many different contexts </b>.
|
195 |
</p>
|
196 |
<p>
|
197 |
+
LLMs are often used to simulate personas and populations, we study the stability and coherence of a simulated <b>population</b> - in contrast to evaluating each persona separately, <b>we evaluate the stability of simulated personas relative to each other, i.e. as a population</b>.
|
198 |
+
We study simulated populations over different contexts, i.e. conversations on different topics.
|
199 |
+
To do that, we leverage the psychological methodology to study the interpersonal stability of personal value expression in a simulated population.
|
200 |
+
We adopt the <b>Schwartz Theory of Basic Personal Values</b>, which defines 10 values: Self-Direction, Stimulation, Hedonism, Achievement, Power, Security, Conformity, Tradition, Benevolence, and Universalism.
|
201 |
+
To evaluate their expression we use the associated questionnaires: <b>PVQ-40</b>, and <b>SVS</b>.
|
202 |
+
</p>
|
203 |
+
<p>
|
204 |
+
The Stick to Your Role! leaderboard aims to provide an up-to-date comparison of recent LLMs based on their ability to coherently simulate popultions.
|
205 |
+
It, in tandem with other minimal-context benchmarks, should enable you to choose the best-suited model for your usecase!
|
206 |
+
If you want to evaluate or add your model you can follow the instructions
|
207 |
+
<a href="{{ url_for('new_model') }}">here</a>.
|
208 |
</p>
|
209 |
</div>
|
210 |
<div class="section">
|
|
|
246 |
<div class="section-title">Validation</div>
|
247 |
<p>
|
248 |
Validity refers to the extent the questionnaire measures what it purports to measure.
|
249 |
+
It can be seen as the questionnaire's accuracy in measuring the intended factors, i.e. values.
|
250 |
Following the recommendations in <a href="https://pubmed.ncbi.nlm.nih.gov/22329443/">this paper</a>,
|
251 |
the validation consists of two phases: Theory-Based Multidimensional Scaling (MDS) and Confirmatory Factor Analysis (CFA).
|
252 |
</p>
|
templates/index.html
CHANGED
@@ -23,13 +23,18 @@
|
|
23 |
box-shadow: 0 4px 8px rgba(0,0,0,0.1);
|
24 |
}
|
25 |
|
26 |
-
h1 {
|
27 |
color: #333;
|
28 |
text-align: center;
|
29 |
}
|
30 |
-
|
31 |
-
|
32 |
-
|
|
|
|
|
|
|
|
|
|
|
33 |
text-align: center;
|
34 |
}
|
35 |
|
@@ -192,20 +197,56 @@
|
|
192 |
text-align: center;
|
193 |
}
|
194 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
195 |
</style>
|
196 |
</head>
|
197 |
<body>
|
198 |
<div class="container">
|
199 |
<h1 class="mt-5">Stick To Your Role! Leaderboard</h1>
|
|
|
|
|
|
|
|
|
|
|
200 |
<p>
|
201 |
The Stick to Your Role! leaderboard compares LLMs based on <b>undesired sensitivity to context change</b>.
|
202 |
-
|
|
|
203 |
As proposed in our <a href="https://arxiv.org/abs/2402.14846">paper</a>,
|
204 |
-
|
205 |
This leaderboard aims to provide such a comparison and extends our paper with a more focused and elaborate experimental setup.
|
206 |
Standard benchmarks present <b>many</b> questions from the <b>same minimal contexts</b> (e.g. multiple choice questions),
|
207 |
we present <b>same</b> questions from <b>many different contexts</b>.
|
208 |
</p>
|
|
|
|
|
|
|
|
|
|
|
209 |
<div class="table-responsive main-table">
|
210 |
<!-- Render the table HTML here -->
|
211 |
{{ main_table_html|safe }}
|
@@ -231,7 +272,7 @@
|
|
231 |
<p>
|
232 |
In addition to Rank-Order stability we compute <b>validity metrics (Stress, Separability, CFI, SRMR, RMSEA)</b>, which are a common practice in psychology.
|
233 |
Validity refers to the extent the questionnaire measures what it purports to measure.
|
234 |
-
It can be seen the questionnaire's accuracy in measuring the intended factors, i.e. values.
|
235 |
For example, basic personal values should be organized in a circular structure, and questions measuring the same value should be correlated.
|
236 |
The table below additionally shows the validity metrics, refer <a href="{{ url_for('about', _anchor='metrics') }}">here</a> for more details.
|
237 |
</p>
|
@@ -245,11 +286,11 @@
|
|
245 |
<li><b>RO-stability</b>: the correlation in the order of simulated participants (ordered based on the expression of the same values) over different contexts</li>
|
246 |
<!--Validation metrics:-->
|
247 |
<li><b>Stress</b>: the MDS fit of the observed value structure to the theoretical circular structure. Stress of 0 indicates 'perfect' fit, 0.025 excellent, 0.05 good, 0.1 fair, and 0.2 poor.</li>
|
248 |
-
<li><b>Separability</b>: the extent to which questions corresponding to different values are linearly separable in the 2D MDS space (linear multi-label classifier accuracy)</li>
|
249 |
-
<li><b>CFI, SRMR, RMSEA</b>: Common Confirmatory Factor Analysis (CFA) metrics showing the fit
|
250 |
<!--Aggregate metrics:-->
|
251 |
<li><b>Ordinal - Win Rate</b>: the score averaged over all metrics (with descending metrics inverted), context pairs (for stability) and contexts (for validity metrics)</li>
|
252 |
-
<li><b>Cardinal - Score</b>: percentage of won games
|
253 |
</ul>
|
254 |
</p>
|
255 |
<div class="table-responsive full-table">
|
@@ -280,7 +321,7 @@
|
|
280 |
</div>
|
281 |
<ul>
|
282 |
<li>Contact: <a href="mailto: [email protected]">[email protected]</a></li>
|
283 |
-
<li>See the <a href="https://sites.google.com/view/llmvaluestability">Project website
|
284 |
<li>See the Flowers team <a href="http://developmentalsystems.org">blog</a> and <a href="https://flowers.inria.fr/">website</a></li>
|
285 |
<li>See Grgur's website and other projects: <a href="https://grgkovac.github.io">https://grgkovac.github.io</a></li>
|
286 |
</ul>
|
|
|
23 |
box-shadow: 0 4px 8px rgba(0,0,0,0.1);
|
24 |
}
|
25 |
|
26 |
+
.container h1 {
|
27 |
color: #333;
|
28 |
text-align: center;
|
29 |
}
|
30 |
+
|
31 |
+
.container h3 {
|
32 |
+
<!-- font-weight: normal;-->
|
33 |
+
<!-- color: #333;-->
|
34 |
+
margin-top: 10px;
|
35 |
+
margin-bottom: 50px;
|
36 |
+
margin-left: 20px;
|
37 |
+
margin-right: 20px;
|
38 |
text-align: center;
|
39 |
}
|
40 |
|
|
|
197 |
text-align: center;
|
198 |
}
|
199 |
|
200 |
+
/* Tooltip styling */
|
201 |
+
.tooltip-inner {
|
202 |
+
max-width: none;
|
203 |
+
}
|
204 |
+
|
205 |
+
th[title] {
|
206 |
+
position: relative;
|
207 |
+
cursor: help;
|
208 |
+
}
|
209 |
+
|
210 |
+
th[title]:hover::after {
|
211 |
+
content: attr(title);
|
212 |
+
position: absolute;
|
213 |
+
bottom: 100%;
|
214 |
+
left: 50%;
|
215 |
+
transform: translateX(-50%);
|
216 |
+
background-color: #333;
|
217 |
+
color: white;
|
218 |
+
padding: 5px 10px;
|
219 |
+
border-radius: 4px;
|
220 |
+
white-space: nowrap;
|
221 |
+
z-index: 1;
|
222 |
+
font-weight: normal;
|
223 |
+
font-size: 14px;
|
224 |
+
}
|
225 |
</style>
|
226 |
</head>
|
227 |
<body>
|
228 |
<div class="container">
|
229 |
<h1 class="mt-5">Stick To Your Role! Leaderboard</h1>
|
230 |
+
<h3>
|
231 |
+
LLMs can role-play different personas by simulating their values and behavior, but can they stick to their role whatever the context?
|
232 |
+
Is simulated Joan of Arc more tradition-driven than Elvis?
|
233 |
+
Will it still be the case after playing chess?
|
234 |
+
</h3>
|
235 |
<p>
|
236 |
The Stick to Your Role! leaderboard compares LLMs based on <b>undesired sensitivity to context change</b>.
|
237 |
+
LLM-exhibited behavior always depends on the context (prompt), while some context-dependence is desired (e.g. following instructions),
|
238 |
+
some is undesired (e.g. drastically changing the simulated value expression based on the interlocutor).
|
239 |
As proposed in our <a href="https://arxiv.org/abs/2402.14846">paper</a>,
|
240 |
+
undesired context-dependence should be seen as a <b>property of LLMs</b> - a dimension of LLM comparison (alongside others such as model size speed or expressed knowledge).
|
241 |
This leaderboard aims to provide such a comparison and extends our paper with a more focused and elaborate experimental setup.
|
242 |
Standard benchmarks present <b>many</b> questions from the <b>same minimal contexts</b> (e.g. multiple choice questions),
|
243 |
we present <b>same</b> questions from <b>many different contexts</b>.
|
244 |
</p>
|
245 |
+
<p>
|
246 |
+
The Stick to You Role! leaderboard focuses on the <b>stability of simulated personal values during role-playing</b>.
|
247 |
+
We study the <b>coherence of a simulated population</b>.
|
248 |
+
In contrast to evaluating each simulated persona separately, we evaluate personas relative to each other, i.e. as a population.
|
249 |
+
</p>
|
250 |
<div class="table-responsive main-table">
|
251 |
<!-- Render the table HTML here -->
|
252 |
{{ main_table_html|safe }}
|
|
|
272 |
<p>
|
273 |
In addition to Rank-Order stability we compute <b>validity metrics (Stress, Separability, CFI, SRMR, RMSEA)</b>, which are a common practice in psychology.
|
274 |
Validity refers to the extent the questionnaire measures what it purports to measure.
|
275 |
+
It can be seen as the questionnaire's accuracy in measuring the intended factors, i.e. values.
|
276 |
For example, basic personal values should be organized in a circular structure, and questions measuring the same value should be correlated.
|
277 |
The table below additionally shows the validity metrics, refer <a href="{{ url_for('about', _anchor='metrics') }}">here</a> for more details.
|
278 |
</p>
|
|
|
286 |
<li><b>RO-stability</b>: the correlation in the order of simulated participants (ordered based on the expression of the same values) over different contexts</li>
|
287 |
<!--Validation metrics:-->
|
288 |
<li><b>Stress</b>: the MDS fit of the observed value structure to the theoretical circular structure. Stress of 0 indicates 'perfect' fit, 0.025 excellent, 0.05 good, 0.1 fair, and 0.2 poor.</li>
|
289 |
+
<li><b>Separability</b>: the extent to which questions corresponding to different values are linearly separable in the 2D MDS space (linear multi-label SVM classifier accuracy)</li>
|
290 |
+
<li><b>CFI, SRMR, RMSEA</b>: Common Confirmatory Factor Analysis (CFA) metrics showing the fit of the posited model of the relation of items (questions) to factors (values) on the observed data, applied here with Magnifying Glass CFA. For CFI >.90 is considered acceptable fit, for SRMR and RMSEA is <.05 considered good fit and <.08 reasonable.</li>
|
291 |
<!--Aggregate metrics:-->
|
292 |
<li><b>Ordinal - Win Rate</b>: the score averaged over all metrics (with descending metrics inverted), context pairs (for stability) and contexts (for validity metrics)</li>
|
293 |
+
<li><b>Cardinal - Score</b>: the percentage of won games, where a game is a comparison each model pair, each metric, and each context pair (for stability) or context (for validity metrics)</li>
|
294 |
</ul>
|
295 |
</p>
|
296 |
<div class="table-responsive full-table">
|
|
|
321 |
</div>
|
322 |
<ul>
|
323 |
<li>Contact: <a href="mailto: [email protected]">[email protected]</a></li>
|
324 |
+
<li>See the <a href="https://sites.google.com/view/llmvaluestability">Project website</a></li>
|
325 |
<li>See the Flowers team <a href="http://developmentalsystems.org">blog</a> and <a href="https://flowers.inria.fr/">website</a></li>
|
326 |
<li>See Grgur's website and other projects: <a href="https://grgkovac.github.io">https://grgkovac.github.io</a></li>
|
327 |
</ul>
|