Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -253,7 +253,7 @@ with gr.Blocks(title=title) as demo:
|
|
253 |
gr.Markdown(" **[Demo is based on sklearn docs found here](https://scikit-learn.org/stable/auto_examples/linear_model/plot_ard.html#sphx-glr-auto-examples-linear-model-plot-ard-py)** <br>")
|
254 |
|
255 |
|
256 |
-
with gr.Tab("
|
257 |
|
258 |
with gr.Row():
|
259 |
n_iter = gr.Slider(value=5, minimum=5, maximum=50, step=1, label="n_iterations")
|
@@ -266,7 +266,7 @@ with gr.Blocks(title=title) as demo:
|
|
266 |
One can observe that with the added noise, none of the models can perfectly recover the coefficients of the original model. All models have more thab 10 non-zero coefficients,
|
267 |
where only 10 are useful. The Bayesian Ridge Regression manages to recover most of the coefficients, while the ARD is more conservative.
|
268 |
""")
|
269 |
-
with gr.Tab("
|
270 |
with gr.Row():
|
271 |
n_iter = gr.Slider(value=5, minimum=5, maximum=50, step=1, label="n_iterations")
|
272 |
btn = gr.Button(value="Plot marginal log likelihoods")
|
@@ -277,7 +277,7 @@ with gr.Blocks(title=title) as demo:
|
|
277 |
Both ARD and Bayesian Ridge minimized the log-likelihood upto an arbitrary cuttoff defined the the n_iter parameter.
|
278 |
"""
|
279 |
)
|
280 |
-
with gr.Tab("
|
281 |
with gr.Row():
|
282 |
degree = gr.Slider(value=5, minimum=5, maximum=50, step=1, label="n_degrees")
|
283 |
btn = gr.Button(value="Plot bayesian regression with polynomial features")
|
|
|
253 |
gr.Markdown(" **[Demo is based on sklearn docs found here](https://scikit-learn.org/stable/auto_examples/linear_model/plot_ard.html#sphx-glr-auto-examples-linear-model-plot-ard-py)** <br>")
|
254 |
|
255 |
|
256 |
+
with gr.Tab("Plot true and estimated coefficients"):
|
257 |
|
258 |
with gr.Row():
|
259 |
n_iter = gr.Slider(value=5, minimum=5, maximum=50, step=1, label="n_iterations")
|
|
|
266 |
One can observe that with the added noise, none of the models can perfectly recover the coefficients of the original model. All models have more thab 10 non-zero coefficients,
|
267 |
where only 10 are useful. The Bayesian Ridge Regression manages to recover most of the coefficients, while the ARD is more conservative.
|
268 |
""")
|
269 |
+
with gr.Tab("Plot marginal log likelihoods"):
|
270 |
with gr.Row():
|
271 |
n_iter = gr.Slider(value=5, minimum=5, maximum=50, step=1, label="n_iterations")
|
272 |
btn = gr.Button(value="Plot marginal log likelihoods")
|
|
|
277 |
Both ARD and Bayesian Ridge minimized the log-likelihood upto an arbitrary cuttoff defined the the n_iter parameter.
|
278 |
"""
|
279 |
)
|
280 |
+
with gr.Tab("Plot bayesian regression with polynomial features"):
|
281 |
with gr.Row():
|
282 |
degree = gr.Slider(value=5, minimum=5, maximum=50, step=1, label="n_degrees")
|
283 |
btn = gr.Button(value="Plot bayesian regression with polynomial features")
|