Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Evaluate Performance and Accuracy of AI Models #297

Closed
jvJUCA opened this issue Jan 31, 2024 · 5 comments
Closed

Evaluate Performance and Accuracy of AI Models #297

jvJUCA opened this issue Jan 31, 2024 · 5 comments
Assignees

Comments

@jvJUCA
Copy link
Member

jvJUCA commented Jan 31, 2024

  • Conduct rigorous evaluations to assess the performance and accuracy of AI models in analyzing user responses compared to human-coded annotations.
  • Measure metrics such as precision, recall, and F1 score to quantify the effectiveness of AI-driven response categorization and sentiment analysis.
  • Benchmark AI models against state-of-the-art approaches and baseline algorithms to identify areas for improvement and optimization.
  • Solicit feedback from researchers and domain experts to validate the relevance and usefulness of AI-generated insights in real-world usability testing scenarios.
@jvJUCA jvJUCA added Documentation Improvements or additions to documentation Question Further information is requested labels Jan 31, 2024
@jvJUCA jvJUCA added this to the [M11] - AI analysis milestone Jan 31, 2024
@kindler-king
Copy link

I would also be interested to take up on this task, given my past experience running A/B tests and evaluation studies on ML models.

I feel that it is also essential to compare the combination of feature engineering and data pre-processing with SOTA ML/NLP techniques to get the best possible result.

The evaluation study needs to be well documented for future work, and can serve as the baseline for choosing models we want to integrate.

I am well connected to researchers being one myself ( check out some of my past work here: https://scholar.google.com/citations?user=vsyHmL0AAAAJ&hl=en)

Would love to discuss this in more detail, and a sample user response or any sample datasets would go a long way in understanding the way forward in going about this study.

@marcgc21
Copy link
Member

@jvJUCA this issue is related to the eye-tracking repository?

@marcgc21 marcgc21 self-assigned this Feb 25, 2024
@kindler-king
Copy link

@marcgc21 Do you know where a good place to start would be?
If I could get my hands on some human-coded annotations and user-response data, then I would have an idea where to start.
Even if we don't have access to data, if I know exactly what kind of data we might need for this comparative analysis, I could look online for some open-access datasets for gold-standard responses and collect user-responses from the community.

Looking forward to some more information on this issue, so that I can start working on it :D
Thanks a lot,

@jvJUCA
Copy link
Member Author

jvJUCA commented Feb 26, 2024

Hi everyone, this issue is related to sentiment analysis, it will be used on the webcam record on a successfully made test.

@kindler-king
Copy link

@jvJUCA That sounds great, do you think we could gain access to a sample video recording as a part of the test set?

I feel for starters we could use a gold standard VIDEO dataset, such as CREMA-D Or RAVDESS as shown here
https://paperswithcode.com/datasets?task=video-emotion-recognition&mod=audio

Then it's a matter of testing the SOTA methods on the dataset of choice and finding the best options, luckily we can get some metrics from the relevant papers citing the dataset and it would be even simpler if they have made their pretrained models available for open access.

I would love to work on this as it's closely related to the other data extraction issue.

@ruxailab ruxailab locked and limited conversation to collaborators Apr 15, 2024
@KarinePistili KarinePistili added Future Work and removed Documentation Improvements or additions to documentation Question Further information is requested labels Apr 15, 2024
@jvJUCA jvJUCA closed this as completed May 29, 2024
@jvJUCA jvJUCA reopened this May 29, 2024
@jvJUCA jvJUCA closed this as completed Jun 19, 2024
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Projects
None yet
Development

No branches or pull requests

4 participants