You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Jun 7, 2023. It is now read-only.
Context: Today in S&T @Carriquiry brought up the fact that a lot of problems CSAFE works on are learning from non-iid data. e.g. compare shoe 1 to shoe 2, then shoe 1 to shoe 3, then shoe 1 to shoe 4, and so on.
Problem: This is a problem because most learning methods treat data as iid. So what can we do to deal with that fact?
So what? Key to the mission of CSAFE: establishing statistical foundations for forensics.
The text was updated successfully, but these errors were encountered:
I stumbled across this paper http://www.jmlr.org/papers/volume11/ralaivola10a/ralaivola10a.pdf, which seeks to provide some probabilistic bounds on the risk for a classifier learned under non iid data. I haven't read it yet, but it may be something to look into. One potential issue with this direction is that I'm pretty sure the theory requires a specific type of loss function and assumes that one is making decisions so as to minimize the expected loss. However, the framework that this paper builds on seems to be a very popular one for understanding the theoretical accuracy properties of machine learning models.
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Context: Today in S&T @Carriquiry brought up the fact that a lot of problems CSAFE works on are learning from non-iid data. e.g. compare shoe 1 to shoe 2, then shoe 1 to shoe 3, then shoe 1 to shoe 4, and so on.
Problem: This is a problem because most learning methods treat data as iid. So what can we do to deal with that fact?
So what? Key to the mission of CSAFE: establishing statistical foundations for forensics.
The text was updated successfully, but these errors were encountered: