[EvalRS 2023][Extended Deadline] [CfP] - Workshop EvalRS 2023. Well-Rounded Recommender Systems For Real-World Deployments

瀏覽次數:4 次
跳到第一則未讀訊息

Claudio Pomo

未讀,
2023年5月7日 中午12:52:172023/5/7
收件者:Machine Learning and Statistics

*** Apologies for cross-posting ***  

  

++ DEADLINE EXTENSION ++


The EvalRS 2023. Well-Rounded Recommender Systems For Real-World Deployments,

in conjunction with the 29TH ACM SIGKDD Conference On Knowledge Discovery And

Data Mining (ACM SIGKDD 2023), ONSITE in Long Beach (CA), USA, 6-10 August 2023.


Full details are available online: https://reclist.io/kdd2023-cup/


# IMPORTANT DATES

Submission deadline: 9 June 2023 AoE (extended)

Notification: 23 June 2023

Camera-ready: 15 July 2023

Workshop date: August 7th


# FORMAT
Research papers and 4 hours hackathon on recSys evaluation. 


The CFP is open (see below); full program and the finalize logistics will be soon announced on the official website: https://reclist.io/kdd2023-cup/ 


# PRIZES

Thanks to Mozilla AI support, we will award monetary prizes for:

  • best paper

  • best student paper

  • best hackathon project


# MOTIVATION & GOALS

EvalRS aims to foster closer partnerships between the academic and industrial sectors regarding the well-rounded evaluation of recommender systems (RS).  


The traditional approach to RS evaluation has been centered on accuracy metrics. However, EvalRS seeks to expand the scope of evaluation techniques beyond just accuracy, to encompass other vital aspects such as fairness, interpretability, and robustness. By bringing together experts from industry, academia, and government, EvalRS creates a forum for discussion and collaboration on the latest trends and challenges across a wide

range of domains. 


EvalRS 2023 follows in the footsteps of the first edition, EvalRS 2022, which featured more than 150 participants and was conducted entirely in the open, with artifacts such as datasets, metrics and evaluation code released back in the community. A review was published in Nature Machine Intelligence (https://www.nature.com/articles/s42256-022-00606-0) emphasizing the first-of-its-kind nature of the workshop, in which theoretical considerations became practical contributions, as participants were asked to “live and breadth” the problem of evaluation through a data challenge. 


We believe that the rounded evaluation of RS is, by nature, a multi-faceted and multi-disciplinary endeavor and that the field as a whole has often been held back by the false dichotomy of

quantitative-and-scalable vs. qualitative-and-manual. The introduction of the hackathon promises to be an additional element of differentiation, making EvalRS 2023 a chance to present cutting-edge work on recommendation systems and network with like-minded, learn-by-doing, and, why not, win a prize!


# TOPICS OF INTEREST

Topics of interest include, but are not limited to:

- Online vs offline evaluation - e.g. making offline evaluation more trustworthy and unbiased;

- Tools and frameworks for the evaluation of RS;

- Empirical studies on the evaluation of RS;

- Reports from real-world deployments - failures, successes, and surprises;

- New metrics and methodologies for evaluation, both quantitative and qualitative;

- Multi-dimensional evaluation, combining multiple recommendation quality factors;

- Multi-disciplinary investigation on ethical questions connected to the deployment and use of RS.


# TYPES OF PAPERS

In EvalRs 2023 We encourage the submission of original contributions along our main topics. Submitted papers will be evaluated (single-blind) according to their originality, technical content, style, clarity, and relevance to the workshop. Papers must be original work and may not be under submission to another venue at the time of review. 


Accepted papers will appear in the workshop proceedings (as we did with EvalRS, we plan on using CEUR for proceedings). 


- Long research/position papers (8 pages, excl references) short research/position papers (4 pages, excl. references), presenting work in

progress, lessons learnt, positions, emerging or future research issues and

directions on Recommender Systems evaluation.


- Extended abstract (2 pages, excl. references), containing descriptions of ongoing

projects or presenting already published results in the area.


Submissions of contributions must be in English, in PDF format in the CEUR-WS two-column conference format available at: http://ceur-ws.org/Vol-XXX/CEURART.zip

or at: https://www.overleaf.com/latex/templates/template-for-submissions-to-ceur-workshop-proceedings-ceur-ws-dot-org/hpvjjzhjxzjk if an Overleaf template is preferred.


# SUBMISSION & PUBLICATION

All papers will undergo a peer review process by at least two expert reviewers to ensure a high standard of quality. Referees will consider originality, significance, technical soundness, clarity of exposition, and relevance to the workshop’s topics.


Research papers should be submitted electronically as a single PDF file through the CMT submission system at the following link https://cmt3.research.microsoft.com/EvalRS2023.


We plan to award monetary prizes to students and participants for outstanding paper contribution.


# ORGANIZING COMMITTEE

Federico Bianchi - Stanford

Patrick John Chia - Coveo

Ciro Greco - Bauplan

Gabriel Moreira - NVIDIA

Claudio Pomo - Politecnico di Bari

Davide Eynard - Mozilla AI

Fahd Husain - Mozilla AI

Jacopo Tagliabue - NYU, Bauplan


回覆所有人
回覆作者
轉寄
0 則新訊息