1st UniDive Training Summer School 2024
Dates: *8 — 12 July 2024*
Location: *Technical University of Moldova*, Chișinău, Moldova
Coordinating Project: UNIDIVE
<https://unidive.lisn.upsaclay.fr/doku.php?id=start> (Universality,
Diversity and Idiosyncrasy in Language Technology)
Website:
https://unidive.lisn.upsaclay.fr/doku.php?id=meetings:other-events:1st_unid…
Cost: *Participants selected on the basis of their application will be
reimbursed, details are below.*
*Apply by:* *May 01, 2024*
CALL FOR APPLICATIONS
We are happy to announce the 1st edition of UNIDIVE Summer School on
Universality, Diversity and Idiosyncrasy in Language Technology. It is
dedicated mainly (but not exclusively) to young researchers and
investigators. Researchers working on low-resourced languages, dialects and
varieties are particularly welcome
SUMMER SCHOOL ACTIVITIES
- Annotation of Universal Dependencies treebank for a new language - a
course by Sylvain Kahane (Université Paris Nanterre, France) and Francis
Tyers (Indiana University, USA)
- Annotation of multiword expressions in a new language - course by
Verginica Mititelu (Romanian Academy) and Voula Giouli (Aristotle
University of Thessaloniki and ILSP, ATHENA RC, Greece)
- Corpus annotation infrastructure (annotation platforms, format
validators, Git etc.) - a course by Daniel Zeman (Charles University,
Czechia), Bruno Guillaume (LORIA, France) and Agata Savary (Université
Paris-Saclay, France)
- A brainstorming hackathon on topics submitted by the trainees
- Poster sessions
APPLICATIONS AND SUBMISSION GUIDELINES
Each applicant should *submit a project* for a construction of a resource
related to the topics of the training school (e.g. a new/enhanced UD
treebank, a new PARSEME corpus, a resource adding a new annotation layer on
top of a UD/PARSEME corpus, etc.). The length of the application should be
2 pages (excluding references). The application should contain:
- The title
- Applicant’s name and affiliation (including the country)
- A list of 3-4 key-words
- Description of a resource related to the topics of the training
school
- Explanation how the participation in the training school will be
useful for the project
- Open questions related to the project which could be addressed
during the brainstorming hackathon
- Short statement of the project phase (planning, started, in the
process of creation)
The projects are to be submitted via the OpenReview
<https://openreview.net/group?id=UniDive/2024/Training_School> portal.
TRAINEE’S SELECTION CRITERIA
We can fund at least 40 trainees, the selection criteria include:
- Trainee’s country: trainees only from COST countries[1]
<#m_5787468061823227072_m_-5162026155145348429__ftn1> and Near-Neighbour
Countries can be funded. See here <https://www.cost.eu/about/members/> and
here <https://www.cost.eu/about/strategy/international-collaboration/>.
- Age: Young Researchers and Investigators, i.e. under the age of 40,
are promoted
- Gender and geographical balance (notably between Inclusiveness
Target Countries and others COST countries)
- Relevance and quality of the project submitted by the trainee
- Status of the language on which the trainee intends to work
(low-resourced languages, dialects or varieties are promoted)
*If you are not selected on the basis of these criteria and you can find
other financial sources to cover your travel, accommodation and meals, you
are also welcome to participate. *
*The authors of the selected projects may optionally present them in a
poster session during the Training School. *
IMPORTANT DATES
Deadline for project submission: May 01, 2024
Notification of acceptance: May 15, 2024
Summer school: July 8-12, 2024
For any inquiry, please contact the organisers at:
victoria.bobicev(a)ia.utm.md
Looking forward to seeing you in Moldova,
Organizing Committee
------------------------------
[1] <#m_5787468061823227072_m_-5162026155145348429__ftnref1> COST
members include 3 categories: COST Full Members, COST Cooperating Members
(Israel), COST Partner Members (South Africa).
Dear KI 2024 Conference Participants,
as the 47th German Conference on Artificial Intelligence (KI 2024)
approaches, we have some exciting news to share!
Extension of Abstract and Full Paper Submission
Due to overwhelming interest and numerous requests, we are pleased to
announce that the deadline for abstract and full paper submissions has been
extended. This extension provides you with additional time to refine your
research and contribute to the vibrant discussions at KI 2024.
Here are the revised submission deadlines:
• Abstract Submission: New deadline is *May 3rd*, 2024
• Full Paper Submission: You now have until *May 12th*, 2024 to submit
your full papers.
Summary of Already Accepted Papers at other AI-Conferences as Extended
Abstracts
We are delighted to inform you that the deadline for submitting summaries
of papers already accepted at other AI conferences, as extended abstracts,
has been set for *June 2nd*.
We encourage all researchers, academics, and industry professionals
interested in artificial intelligence to take advantage of this
opportunity. KI 2024 is a premier German AI conference that brings together
experts from various domains, fostering collaboration and knowledge
exchange.
For more details about the conference program, venue, and submission
guidelines, please visit the official KI 2024 website
<https://www.informatik.uni-wuerzburg.de/ki24/> (
https://www.informatik.uni-wuerzburg.de/ki24/).
Thank you for your continued support, and we look forward to your valuable
contributions at KI 2024!
Best regards,
KI 2024 Team
The first workshop on evaluating IR systems with Large Language Models
(LLMs) is accepting submissions that describe original research findings,
preliminary research results, proposals for new work, and recent relevant
studies already published in high-quality venues.
Topics of interest
We welcome both full papers and extended abstract submissions on the
following topics, including but not limited to:
- LLM-based evaluation metrics for traditional IR and generative IR.
- Agreement between human and LLM labels.
- Effectiveness and/or efficiency of LLMs to produce robust relevance
labels.
- Investigating LLM-based relevance estimators for potential systemic
biases.
- Automated evaluation of text generation systems.
- End-to-end evaluation of Retrieval Augmented Generation systems.
- Trustworthiness in the world of LLMs evaluation.
- Prompt engineering in LLMs evaluation.
- Effectiveness and/or efficiency of LLMs as ranking models.
- LLMs in specific IR tasks such as personalized search, conversational
search, and multimodal retrieval.
- Challenges and future directions in LLM-based IR evaluation.
Submission guidelines
We welcome the following submissions:
- Previously unpublished manuscripts will be accepted as extended
abstracts and full papers (any length between 1 - 9 pages) with unlimited
references, formatted according to the latest ACM SIG proceedings template
available at http://www.acm.org/publications/proceedings-template.
- Published manuscripts can be submitted in their original format.
All submissions should be made through Easychair:
https://easychair.org/conferences/?conf=llm4eval
All papers will be peer-reviewed (single-blind) by the program committee
and judged by their relevance to the workshop, especially to the main
themes identified above, and their potential to generate discussion. For
already published studies, the paper can be submitted in the original
format. These submissions will be reviewed for their relevance to this
workshop. All submissions must be in English (PDF format).
All accepted papers will have a poster presentation with a few selected for
spotlight talks. Accepted papers may be uploaded to arXiv.org, allowing
submission elsewhere as they will be considered non-archival. The
workshop’s website will maintain a link to the arXiv versions of the papers.
Important Dates
- Submission Deadline: April 25th May 2nd, 2024 (AoE time)
- Acceptance Notifications: May 31st, 2024 (AoE time)
- Workshop date: July 18, 2024
Website
For more information, visit the workshop website:
https://llm4eval.github.io/
Contact
For any questions about paper submission, you may contact the workshop
organizers at llm4eval(a)easychair.org
International Conference ‘New Trends in Translation and Technology’ (NeTTT’2024)
Varna, Bulgaria, 3-6 July 2024
https://nettt-conference.com/
*** Last Call for Papers (Submission deadline approaching: 30 April 2024) ***
# The conference
The second edition of the forthcoming International Conference ‘New Trends in Translation and Technology’ (NeTTT’2024) will take place in Varna, Bulgaria, 3-6 July 2024.
Continuing the tradition of the first edition of the NeTTT conference and HiT-IT events series, the objective of the conference is (i) to bridge the gap between academia and industry in the field of translation and interpreting by bringing together academics in linguistics, translation studies, machine translation and natural language processing, developers, practitioners, language service providers and vendors who work on or are interested in different aspects of technology for translation and interpreting, and (ii) to be a distinctive event for discussing the latest developments and practices. NeTTT’2024 invites all professionals who would like to learn about the new trends, present the latest work or/and share their experience in the field, and who would like to establish business and research contacts, collaborations and new ventures.
The conference will take the form of presentations (peer-reviewed research and user presentations, keynote speeches), and posters; it will also feature panel discussions. The accepted papers will be published as open-access conference e-proceedings.
# Conference topics
Contributions are invited on any topic related to latest technology and practices in machine translation, translation, subtitling, localisation and interpreting. NeTTT’2024 will feature a Special Theme Track "Future of Translation Technology in the Era of LLMs and Generative AI".
The conference topics include but are not limited to:
## CAT tools
- Translation Memory (TM) systems
- NLP and MT for translation memory systems
- Terminology extraction tools
- Localisation tools
## Machine Translation
- Latest developments in Neural Machine Translation
- MT for under-resourced languages
- MT with low computing resources
- Multimodal MT
- Integration of MT in TM systems
- Resources for MT
## Technologies for MT deployment
- MT evaluation techniques, metrics and evaluation results
- Human evaluations of MT output
- Evaluating MT in a real-world setting
- Quality estimation for MT
- Domain adaptation
## Translation Studies
- Corpus-based studies applied to translation
- Corpora and resources for translation
- Translationese
- Cognitive effort and eye-tracking experiments in translation
## Interpreting studies
- Corpus-based studies applied to interpreting
- Corpora and resources for interpreting
- Interpretese
- Resources for interpreting and interpreting technology applications
- Cognitive effort and eye-tracking experiments in interpreting
## Interpreting technology
- Machine interpreting
- Computer-aided interpreting
- NLP for dialogue interpreting
- Development of NLP based applications for communication in public service settings (healthcare, education, law, emergency services)
## Emerging Areas in Translation and Interpreting
- MT and translation tools for literary texts and creative texts
- MT for social media and real-time conversations
- Sign language recognition and translation
## Subtitling
- NLP and MT for subtitling
- Latest technology for subtitling
## User needs
- Analysis of translators’ and interpreters’ needs in terms of translation and interpreting technology
- User requirements for interpreting and translation tools
- Incorporating human knowledge into translation and interpreting technology
- What existing translators’ (including subtitlers’) and interpreters’ tools do not offer
- User requirements for electronic resources for translators and interpreters
- Translation and interpreting workflows in larger organisations and the tools for translation and interpreting employed
## The business of translation and interpreting
- Translation workflow and management
- Technology adoption by translators and industry
- Setting up translation /interpreting / language provider company
## Teaching translation and interpreting
- Teaching Machine Translation
- Teaching translation technology
- Teaching interpreting technology
- Latest AI developments in the syllabi of translation and interpreting curricula
## Ethical issues in translation and technology
- Bias and fairness in MT
- Privacy and security in cloud MT systems
- Transparency and explainability of MT systems
- Environmental impact on MT systems
# Special Theme Track - Future of Translation Technology in the Era of LLMs and Generative AI
We are excited to share that NeTTT’2024 will have a special theme with the goal of stimulating discussion around Large Language Models, Generative AI and the Future of Translation and Interpreting Technology. While the new generation of Large Language Models such as CHATGPT and LLAMA showcase remarkable advancements in language generation and understanding, we find ourselves in uncharted territory when it comes to their performance on various Translation and Interpreting Technology tasks with regards to fairness, interpretability, ethics and transparency.
The theme track invites studies on how LLMs perform on Translation and Interpreting Technology tasks and applications, and what this means for the future of the field. The possible topics of discussion include (but are not limited to) the following:
- Changes in the translators and interpreters’ professions in the new AI era especially as a result of the latest developments in LLMSs and Generative AI
- Generative AI and translation
- Generative AI and interpreting
- Augmenting machine translation systems with generative AI
- Domain and terminology adaptation with Large Language Models
- Literary translation with Large Language Models
- Improving Machine Translation Quality with Contextual Prompts in Large Language Models
- Prompt engineering for translation
- Generative AI for professional translation
- Generative AI for professional interpreting
# Keynote speakers
We are delighted to announce the NeTTT’2024 keynote speakers
- Helena Moniz (University of Lisbon and Unbabel), President of the European Association of Machine Translation
- Carla Parra Escartín (RWS Language Weaver)
# Tutorial (3 July 2024)
- Tharindu Ranasinghe (Aston University), Quality Estimation for Machine Translation
# Programme Committee
The Programme Committee of NeTTT’2024 is listed https://nettt-conference.com/26844-2/.
# Conference Chairs
- Ruslan Mitkov (Lancaster University)
- Gloria Corpas Pastor (University of Malaga)
# Programme Chairs
- Constantin Orasan (University of Surrey)
- Tharindu Ranasinghe (Aston University)
# Sponsorship Chair
- Vilelmini Sosoni (Ionian University)
# Publication Chair
- Maria Kunilovskaya (University of Saarland)
# Organising Committee
- Organising Committee of NeTTT’2024 is listed https://nettt-conference.com/organisers/
# Submissions and publication
NETTT’2024 invites the following types of submissions:
User papers – for industry and practitioners. References to related work are optional. Allowed paper length: between 1 and 4 pages.
Academic submissions, in three different categories (have to follow formatting requirements, references to related work are required):
• (academic) full papers – describing original completed research. Allowed paper length: maximum 12 pages + unlimited references.
• (academic) work-in-progress papers/posters – describing work in progress, late breaking research, papers at a more conceptual stage, and other types of papers that do not fit in the ‘full’ papers category. Allowed paper length: maximum 7 pages + unlimited references.
• (academic) demo papers – describing working systems. Allowed paper length: maximum 5 pages + unlimited references. In addition to the papers, the authors will be expected to demonstrate the systems at the workshop.
The conference will not consider and evaluate abstracts only.
Each submission will be reviewed by three members of the Programme Committee. Submission is organised via Softconf START conference management system at https://softconf.com/n/nettt2024.
For submitting the papers, we invite the authors to comply with the Springer format, following the templates:
• LaTeX: https://resource-cms.springernature.com/springer-cms/rest/v1/content/192386…,
• Overleaf: https://www.overleaf.com/latex/templates/springer-lecture-notes-in-computer…,
• Word: https://resource-cms.springernature.com/springer-cms/rest/v1/content/192387….
The accepted papers will be published in the conference proceedings and made available online on the conference website. Authors of accepted papers will receive guidelines regarding how to produce camera-ready versions of their papers.
The final version of the accepted papers will be published in e-proceedings with assigned ISBN and DOI.
All accepted papers will be included in the conference e-proceedings which will be available at the conference website.
# Schedule
- Submission deadline: 30 April 2024
- Notification: 5 June 2024
- Final version due: 20 June 2024
All deadlines are valid for 23.59 Anywhere on Earth.
# Registration
Conference registration is open on https://nettt-conference.com/fees-registration/
The promotional early registration fee has been extended to 17 March 2024.
# Venue
The conference will take place at https://www.chernomorebg.com/en/conference-centre.html, Varna, situated only 200 m away from the fine sandy Black Sea beach.
# Sponsors
We are proud to announce the conference sponsors:
OONA - Diamond Sponsor
Pangeanic – Gold Sponsor
MITRA Translations – Silver Sponsor
Juremy – Bronze Sponsor
# Further information and contact details
The conference website is https://nettt-conference.com and will be updated on a regular basis. For further information, please contact us at nettt2024(a)nettt-conference.com
Dr Tharindu Ranasinghe
Lecturer in Computer Science
School of Informatics and Digital Engineering
Birmingham, B4 7ET, UK
aston.ac.uk
The first workshop on evaluating IR systems with Large Language Models
(LLMs) is accepting submissions that describe original research findings,
preliminary research results, proposals for new work, and recent relevant
studies already published in high-quality venues.
Topics of interest
We welcome both full papers and extended abstract submissions on the
following topics, including but not limited to:
- LLM-based evaluation metrics for traditional IR and generative IR.
- Agreement between human and LLM labels.
- Effectiveness and/or efficiency of LLMs to produce robust relevance
labels.
- Investigating LLM-based relevance estimators for potential systemic
biases.
- Automated evaluation of text generation systems.
- End-to-end evaluation of Retrieval Augmented Generation systems.
- Trustworthiness in the world of LLMs evaluation.
- Prompt engineering in LLMs evaluation.
- Effectiveness and/or efficiency of LLMs as ranking models.
- LLMs in specific IR tasks such as personalized search, conversational
search, and multimodal retrieval.
- Challenges and future directions in LLM-based IR evaluation.
Submission guidelines
We welcome the following submissions:
- Previously unpublished manuscripts will be accepted as extended
abstracts and full papers (any length between 1 - 9 pages) with unlimited
references, formatted according to the latest ACM SIG proceedings template
available at http://www.acm.org/publications/proceedings-template.
- Published manuscripts can be submitted in their original format.
All submissions should be made through Easychair:
https://easychair.org/conferences/?conf=llm4eval
All papers will be peer-reviewed (single-blind) by the program committee
and judged by their relevance to the workshop, especially to the main
themes identified above, and their potential to generate discussion. For
already published studies, the paper can be submitted in the original
format. These submissions will be reviewed for their relevance to this
workshop. All submissions must be in English (PDF format).
All accepted papers will have a poster presentation with a few selected for
spotlight talks. Accepted papers may be uploaded to arXiv.org, allowing
submission elsewhere as they will be considered non-archival. The
workshop’s website will maintain a link to the arXiv versions of the papers.
Important Dates
- Submission Deadline: April 25th, 2024 (AoE time)
- Acceptance Notifications: May 31st, 2024 (AoE time)
- Workshop date: July 18, 2024
Website
For more information, visit the workshop website:
https://llm4eval.github.io/
Contact
For any questions about paper submission, you may contact the workshop
organizers at llm4eval(a)easychair.org
*1st UniDive training Summer School 2024*
Dates: *8 — 12 July 2024*
Location: *Technical University of Moldova*, Chișinău, Moldova
Coordinating Project: UNIDIVE
<https://unidive.lisn.upsaclay.fr/doku.php?id=start> (Universality,
Diversity and Idiosyncrasy in Language Technology)
Website:
https://unidive.lisn.upsaclay.fr/doku.php?id=meetings:other-events:1st_unid…
Cost: *Participants selected on the basis of their application will be
reimbursed, details are below.*
*Apply by:* *April 29, 2024*
CALL FOR APPLICATIONS
We are happy to announce the 1th edition of UNIDIVE Summer School on
Universality, Diversity and Idiosyncrasy in Language Technology. It is
dedicated mainly (but not exclusively) to young researchers and
investigators. Researchers working on low-resourced languages, dialects and
varieties are particularly welcome
SUMMER SCHOOL ACTIVITIES
- Annotation of Universal Dependencies treebank for a new language - a
course by Sylvain Kahane (Université Paris Nanterre, France) and Francis
Tyers (Indiana University, USA)
- Annotation of multiword expressions in a new language - course by
Verginica Mititelu (Romanian Academy) and Voula Giouli (Aristotle
University of Thessaloniki and ILSP, ATHENA RC, Greece)
- Corpus annotation infrastructure (annotation platforms, format
validators, Git etc.) - a course by Daniel Zeman (Charles University,
Czechia), Bruno Guillaume (LORIA, France) and Agata Savary (Université
Paris-Saclay, France)
- A brainstorming hackathon on topics submitted by the trainees
- Poster sessions
APPLICATIONS AND SUBMISSION GUIDELINES
Each applicant should *submit a project* for a construction of a resource
related to the topics of the training school (e.g. a new/enhanced UD
treebank, a new PARSEME corpus, a resource adding a new annotation layer on
top of a UD/PARSEME corpus, etc.). The length of the application should be
2 pages (excluding references). The application should contain:
- The title
- Applicant’s name and affiliation (including the country)
- A list of 3-4 key-words
- Description of a resource related to the topics of the training
school
- Explanation how the participation in the training school will be
useful for the project
- Open questions related to the project which could be addressed
during the brainstorming hackathon
- Short statement of the project phase (planning, started, in the
process of creation)
The projects are to be submitted via the OpenReview
<https://openreview.net/group?id=UniDive/2024/Training_School> portal.
TRAINEE’S SELECTION CRITERIA
We can fund at least 40 trainees, the selection criteria include:
- Trainee’s country: trainees only from COST countries[1]
<#m_-2062973076706449294_m_-5162026155145348429__ftn1> and Near-Neighbour
Countries can be funded. See here <https://www.cost.eu/about/members/> and
here <https://www.cost.eu/about/strategy/international-collaboration/>.
- Age: Young Researchers and Investigators, i.e. under the age of 40,
are promoted
- Gender and geographical balance (notably between Inclusiveness
Target Countries and others COST countries)
- Relevance and quality of the project submitted by the trainee
- Status of the language on which the trainee intends to work
(low-resourced languages, dialects or varieties are promoted)
*If you are not selected on the basis of these criteria and you can find
other financial sources to cover your travel, accommodation and meals, you
are also welcome to participate. *
*The authors of the selected projects may optionally present them in a
poster session during the Training School. *
IMPORTANT DATES
Deadline for project submission: April 29, 2024
Notification of acceptance: May 15, 2024
Summer school: July 8-12, 2024
For any inquiry, please contact the organisers at:
victoria.bobicev(a)ia.utm.md
Looking forward to seeing you in Moldova,
Organizing committee
------------------------------
[1] <#m_-2062973076706449294_m_-5162026155145348429__ftnref1> COST
members include 3 categories: COST Full Members, COST Cooperating Members
(Israel), COST Partner Members (South Africa).
Overview
The first workshop on evaluating IR systems with Large Language Models
(LLMs) is accepting submissions that describe original research findings,
preliminary research results, proposals for new work, and recent relevant
studies already published in high-quality venues. The workshop will have
both an in-person and virtual component, and submissions are welcome even
for researchers who cannot attend in person, as they will present their
work in the virtual component.
Topics of interest
We welcome both full papers and extended abstract submissions on the
following topics, including but not limited to:
- LLM-based evaluation metrics for traditional IR and generative IR.
- Agreement between human and LLM labels.
- Effectiveness and/or efficiency of LLMs to produce robust relevance
labels.
- Investigating LLM-based relevance estimators for potential systemic
biases.
- Automated evaluation of text generation systems.
- End-to-end evaluation of Retrieval Augmented Generation systems.
- Trustworthiness in the world of LLMs evaluation.
- Prompt engineering in LLMs evaluation.
- Effectiveness and/or efficiency of LLMs as ranking models.
- LLMs in specific IR tasks such as personalized search, conversational
search, and multimodal retrieval.
- Challenges and future directions in LLM-based IR evaluation.
Submission guidelines
We welcome the following submissions:
- Previously unpublished manuscripts will be accepted as extended
abstracts and full papers (any length between 1 - 9 pages) with unlimited
references, formatted according to the latest ACM SIG proceedings template
available at http://www.acm.org/publications/proceedings-template.
- Published manuscripts can be submitted in their original format.
All submissions should be made through Easychair:
https://easychair.org/conferences/?conf=llm4eval
All papers will be peer-reviewed (single-blind) by the program committee
and judged by their relevance to the workshop, especially to the main
themes identified above, and their potential to generate discussion. For
already published studies, the paper can be submitted in the original
format. These submissions will be reviewed for their relevance to this
workshop. All submissions must be in English (PDF format).
Please note the workshop will have an in-person (to be held with SIGIR
2024) and virtual component (to be held at a later date on SIGIR VF).
During submission, the authors should select their preferred component. All
accepted papers will have a poster presentation with a few selected for
spotlight talks. Accepted papers may be uploaded to arXiv.org, allowing
submission elsewhere as they will be considered non-archival. The
workshop’s website will maintain a link to the arXiv versions of the papers.
Important Dates
- Submission Deadline: April 25th, 2024 (AoE time)
- Acceptance Notifications: May 31st, 2024 (AoE time)
- Workshop date: July 18, 2024
Website and Contact
More details are available at https://llm4eval.github.io/cfp/.
For any questions about paper submission, you may contact the workshop
organizers at llm4eval(a)easychair.org
RispondiInoltra
International Conference ‘New Trends in Translation and Technology’ (NeTTT’2024)
Varna, Bulgaria, 3-6 July 2024
https://nettt-conference.com/
*** Submission Deadline Extended to 30 April 2024 ***
# The conference
The second edition of the forthcoming International Conference ‘New Trends in Translation and Technology’ (NeTTT’2024) will take place in Varna, Bulgaria, 3-6 July 2024.
Continuing the tradition of the first edition of the NeTTT conference and HiT-IT events series, the objective of the conference is (i) to bridge the gap between academia and industry in the field of translation and interpreting by bringing together academics in linguistics, translation studies, machine translation and natural language processing, developers, practitioners, language service providers and vendors who work on or are interested in different aspects of technology for translation and interpreting, and (ii) to be a distinctive event for discussing the latest developments and practices. NeTTT’2024 invites all professionals who would like to learn about the new trends, present the latest work or/and share their experience in the field, and who would like to establish business and research contacts, collaborations and new ventures.
The conference will take the form of presentations (peer-reviewed research and user presentations, keynote speeches), and posters; it will also feature panel discussions. The accepted papers will be published as open-access conference e-proceedings.
# Conference topics
Contributions are invited on any topic related to latest technology and practices in machine translation, translation, subtitling, localisation and interpreting. NeTTT’2024 will feature a Special Theme Track "Future of Translation Technology in the Era of LLMs and Generative AI".
The conference topics include but are not limited to:
## CAT tools
- Translation Memory (TM) systems
- NLP and MT for translation memory systems
- Terminology extraction tools
- Localisation tools
## Machine Translation
- Latest developments in Neural Machine Translation
- MT for under-resourced languages
- MT with low computing resources
- Multimodal MT
- Integration of MT in TM systems
- Resources for MT
## Technologies for MT deployment
- MT evaluation techniques, metrics and evaluation results
- Human evaluations of MT output
- Evaluating MT in a real-world setting
- Quality estimation for MT
- Domain adaptation
## Translation Studies
- Corpus-based studies applied to translation
- Corpora and resources for translation
- Translationese
- Cognitive effort and eye-tracking experiments in translation
## Interpreting studies
- Corpus-based studies applied to interpreting
- Corpora and resources for interpreting
- Interpretese
- Resources for interpreting and interpreting technology applications
- Cognitive effort and eye-tracking experiments in interpreting
## Interpreting technology
- Machine interpreting
- Computer-aided interpreting
- NLP for dialogue interpreting
- Development of NLP based applications for communication in public service settings (healthcare, education, law, emergency services)
## Emerging Areas in Translation and Interpreting
- MT and translation tools for literary texts and creative texts
- MT for social media and real-time conversations
- Sign language recognition and translation
## Subtitling
- NLP and MT for subtitling
- Latest technology for subtitling
## User needs
- Analysis of translators’ and interpreters’ needs in terms of translation and interpreting technology
- User requirements for interpreting and translation tools
- Incorporating human knowledge into translation and interpreting technology
- What existing translators’ (including subtitlers’) and interpreters’ tools do not offer
- User requirements for electronic resources for translators and interpreters
- Translation and interpreting workflows in larger organisations and the tools for translation and interpreting employed
## The business of translation and interpreting
- Translation workflow and management
- Technology adoption by translators and industry
- Setting up translation /interpreting / language provider company
## Teaching translation and interpreting
- Teaching Machine Translation
- Teaching translation technology
- Teaching interpreting technology
- Latest AI developments in the syllabi of translation and interpreting curricula
## Ethical issues in translation and technology
- Bias and fairness in MT
- Privacy and security in cloud MT systems
- Transparency and explainability of MT systems
- Environmental impact on MT systems
# Special Theme Track - Future of Translation Technology in the Era of LLMs and Generative AI
We are excited to share that NeTTT’2024 will have a special theme with the goal of stimulating discussion around Large Language Models, Generative AI and the Future of Translation and Interpreting Technology. While the new generation of Large Language Models such as CHATGPT and LLAMA showcase remarkable advancements in language generation and understanding, we find ourselves in uncharted territory when it comes to their performance on various Translation and Interpreting Technology tasks with regards to fairness, interpretability, ethics and transparency.
The theme track invites studies on how LLMs perform on Translation and Interpreting Technology tasks and applications, and what this means for the future of the field. The possible topics of discussion include (but are not limited to) the following:
- Changes in the translators and interpreters’ professions in the new AI era especially as a result of the latest developments in LLMSs and Generative AI
- Generative AI and translation
- Generative AI and interpreting
- Augmenting machine translation systems with generative AI
- Domain and terminology adaptation with Large Language Models
- Literary translation with Large Language Models
- Improving Machine Translation Quality with Contextual Prompts in Large Language Models
- Prompt engineering for translation
- Generative AI for professional translation
- Generative AI for professional interpreting
# Keynote speakers
We are delighted to announce the NeTTT’2024 keynote speakers
- Helena Moniz (University of Lisbon and Unbabel), President of the European Association of Machine Translation
- Carla Parra Escartín (RWS Language Weaver)
# Tutorial (3 July 2024)
- Tharindu Ranasinghe (Aston University), Quality Estimation for Machine Translation
# Programme Committee
The Programme Committee of NeTTT’2024 is listed https://nettt-conference.com/26844-2/.
# Conference Chairs
- Ruslan Mitkov (Lancaster University)
- Gloria Corpas Pastor (University of Malaga)
# Programme Chairs
- Constantin Orasan (University of Surrey)
- Tharindu Ranasinghe (Aston University)
# Sponsorship Chair
- Vilelmini Sosoni (Ionian University)
# Publication Chair
- Maria Kunilovskaya (University of Saarland)
# Organising Committee
- Organising Committee of NeTTT’2024 is listed https://nettt-conference.com/organisers/
# Submissions and publication
NETTT’2024 invites the following types of submissions:
User papers – for industry and practitioners. References to related work are optional. Allowed paper length: between 1 and 4 pages.
Academic submissions, in three different categories (have to follow formatting requirements, references to related work are required):
• (academic) full papers – describing original completed research. Allowed paper length: maximum 12 pages + unlimited references.
• (academic) work-in-progress papers/posters – describing work in progress, late breaking research, papers at a more conceptual stage, and other types of papers that do not fit in the ‘full’ papers category. Allowed paper length: maximum 7 pages + unlimited references.
• (academic) demo papers – describing working systems. Allowed paper length: maximum 5 pages + unlimited references. In addition to the papers, the authors will be expected to demonstrate the systems at the workshop.
The conference will not consider and evaluate abstracts only.
Each submission will be reviewed by three members of the Programme Committee. Submission is organised via Softconf START conference management system at https://softconf.com/n/nettt2024.
For submitting the papers, we invite the authors to comply with the Springer format, following the templates:
• LaTeX: https://resource-cms.springernature.com/springer-cms/rest/v1/content/192386…,
• Overleaf: https://www.overleaf.com/latex/templates/springer-lecture-notes-in-computer…,
• Word: https://resource-cms.springernature.com/springer-cms/rest/v1/content/192387….
The accepted papers will be published in the conference proceedings and made available online on the conference website. Authors of accepted papers will receive guidelines regarding how to produce camera-ready versions of their papers.
The final version of the accepted papers will be published in e-proceedings with assigned ISBN and DOI.
All accepted papers will be included in the conference e-proceedings which will be available at the conference website.
# Schedule
- Submission deadline: 30 April 2024
- Notification: 5 June 2024
- Final version due: 20 June 2024
All deadlines are valid for 23.59 Anywhere on Earth.
# Registration
Conference registration is open on https://nettt-conference.com/fees-registration/
The promotional early registration fee has been extended to 17 March 2024.
# Venue
The conference will take place at https://www.chernomorebg.com/en/conference-centre.html, Varna, situated only 200 m away from the fine sandy Black Sea beach.
# Further information and contact details
The conference website is https://nettt-conference.com and will be updated on a regular basis. For further information, please contact us at nettt2024(a)nettt-conference.com
Dr Tharindu Ranasinghe
Lecturer in Computer Science
School of Informatics and Digital Engineering
Birmingham, B4 7ET, UK
aston.ac.uk
We invite you to participate in the shared task on Explainability of Cross-lingual Emotion Detection (EXALT), organized as part of WASSA 2024 at ACL 2024. The goal of this task is two-fold: developing highly performant models for cross-lingual emotion detection while also encouraging model explainability and interpretability.
**Task Description**
You can participate in three different tracks:
* Track 1: Cross-Lingual Emotion Detection, which involves emotion classification with 5 labels with transfer from English to a multi-lingual test set that also includes French, Spanish, Dutch and Russian tweets.
* Track 2: Trigger Word detection (binary), a task where participants have to identify which words in a text are most essential for expressing emotion. For this track, we treat trigger word detection as a binary classification task.
* Track 3: Trigger Word detection (numerical), a task where participants have to identify which words in a text are most essential for expressing emotion. For this track, participants can provide numerical importance scores for each individual word.
For more information, please refer to the web page:
https://lt3.ugent.be/exalt/
The data is available through: https://codalab.lisn.upsaclay.fr/competitions/17730<https://codalab.lisn.upsaclay.fr/competitions/17730#learn_the_details-overv…>
**Important Dates**
* 15 February, 2024: first call for participation. Training and validation data released, CodaLab opens for team registrations.
* 13 May, 2024: Evaluation period/phase begins (= test data released)
* 15 May, 2024: Evaluation period/phase ends (= system submission deadline)
* 17 May, 2024: Announcement of final results
* 22 May, 2024: Submission of system description papers for review (we encourage participants to share models, code, fact sheets, extra data, etc. with the community through github or other repositories on paper publication.)
* 22 June, 2024: Notification of Acceptance
* 01 July, 2024: Camera-ready deadline
* 15-16 August, 2024: Presentations of the systems at the WASSA 2024 workshop (Bangkok, Thailand)
**Task Organizers**
Aaron Maladry
<https://lt3.ugent.be/people/aaron-maladry/>LT3, Ghent University
Pranaydeep Singh
<https://lt3.ugent.be/people/pranaydeep-singh/>LT3, Ghent University
Els Lefever
<https://lt3.ugent.be/people/els-lefever/>LT3, Ghent University
**Contact**
You can contact all organizers of the shared task at exalt-wassa2024(a)googlegroups.com<mailto:exalt-wassa2024@googlegroups.com>
We invite proposals for tasks to be run as part of SemEval-2025
<https://semeval.github.io/SemEval2025/>. SemEval (the International
Workshop on Semantic Evaluation) is an ongoing series of evaluations of
computational semantics systems, organized under the umbrella of SIGLEX
<https://siglex.org/>, the Special Interest Group on the Lexicon of the
Association for Computational Linguistics.
SemEval tasks explore the nature of meaning in natural languages: how to
characterize meaning and how to compute it. This is achieved in practical
terms, using shared datasets and standardized evaluation metrics to
quantify the strengths and weaknesses and possible solutions. SemEval tasks
encompass a broad range of semantic topics from the lexical level to the
discourse level, including word sense identification, semantic parsing,
coreference resolution, and sentiment analysis, among others.
For SemEval-2025 <https://semeval.github.io/SemEval2025/cft>, we welcome
tasks that can test an automatic system for the semantic analysis of text
(e.g., intrinsic semantic evaluation, or an application-oriented
evaluation). We especially encourage tasks for languages other than
English, cross-lingual tasks, and tasks that develop novel applications of
computational semantics. See the websites of previous editions of SemEval
to get an idea about the range of tasks explored, e.g. SemEval-2020
<http://alt.qcri.org/semeval2020/> and SemEval-2021-/2023/2024
<https://semeval.github.io/>.
We strongly encourage proposals based on pilot studies that have already
generated initial data, evaluation measures and baselines. In this way, we
can avoid unforeseen challenges down the road which that may delay the task.
For example, you may see this task proposal
<https://semeval.github.io/semeval2024_shared_task6_proposals_template.pdf>as
a sample.
In case you are not sure whether a task is suitable for SemEval, please
feel free to get in touch with the SemEval organizers at
semevalorganizers(a)gmail.com to discuss your idea.
=== Task Selection ===
Task proposals will be reviewed by experts, and reviews will serve as the
basis for acceptance decisions. Everything else being equal, more
innovative new tasks will be given preference over task reruns. Task
proposals will be evaluated on:
- Novelty: Is the task on a compelling new problem that has not been
explored much in the community? Is the task a rerun, but covering
substantially new ground (new subtasks, new types of data, new languages,
etc.)?
- Interest: Is the proposed task likely to attract a sufficient number
of participants?
- Data: Are the plans for collecting data convincing? Will the resulting
data be of high quality? Will annotations have meaningfully high
inter-annotator agreements? Have all appropriate licenses for use and
re-use of the data after the evaluation been secured? Have all
international privacy concerns been addressed? Will the data annotation be
ready on time?
- Evaluation: Is the methodology for evaluation sound? Is the necessary
infrastructure available or can it be built in time for the shared task?
Will research inspired by this task be able to evaluate in the same manner
and on the same data after the initial task?
- Impact: What is the expected impact of the data in this task on future
research beyond the SemEval Workshop?
-
Ethical: The data must be compliant with privacy policies. e.g.
a) avoid personally identifiable information (PII). Tasks aimed at
identifying specific people will not be accepted,
b) avoid medical decision making (compliance with HIPAA, do not try to
replace medical professionals, especially if it has anything to do with
mental health)
c) these are representative and not exhaustive
=== New Tasks vs. Task Reruns ===
We welcome both new tasks and task reruns. For a new task, the proposal
should address whether the task would be able to attract participants.
Preference will be given to novel tasks that have not received much
attention yet.
For reruns of previous shared tasks (whether or not the previous task was
part of SemEval), the proposal should address the need for another
iteration of the task. Valid reasons include: a new form of evaluation
(e.g. a new evaluation metric, a new application-oriented scenario), new
genres or domains (e.g. social media, domain-specific corpora), or a
significant expansion in scale. We further discourage carrying over a
previous task and just adding new subtasks, as this can lead to the
accumulation of too many subtasks. Evaluating on a different dataset with
the same task formulation, or evaluating on the same dataset with a
different evaluation metric, typically should not be considered a separate
subtask.
=== Task Organization ===
We welcome people who have never organized a SemEval task before, as well
as those who have. Apart from providing a dataset, task organizers are
expected to:
- Verify the data annotations have sufficient inter-annotator agreement
- Verify licenses for the data allow its use in the competition and
afterwards. In particular, text that is publicly available online is not
necessarily in the public domain; unless a license has been provided, the
author retains all rights associated with their work, including copying,
sharing and publishing. For more information, see:
https://creativecommons.org/faq/#what-is-copyright-and-why-does-it-matter
- Resolve any potential security, privacy, or ethical concerns about the
data
- Commit to make the data available after the task
- Provide task participants with format checkers and standard scorers.
- Provide task participants with baseline systems to use as a starting
point (in order to lower the obstacles to participation). A baseline system
typically contains code that reads the data, creates a baseline response
(e.g. random guessing, majority class prediction), and outputs the
evaluation results. Whenever possible, baseline systems should be written
in widely used programming languages and/or should be implemented as a
component for standard NLP pipelines.
- Create a mailing list and website for the task and post all relevant
information there.
- Create a CodaLab or other similar competition for the task and upload
the evaluation script.
- Manage submissions on CodaLab or a similar competition site.
- Write a task description paper to be included in SemEval proceedings,
and present it at the workshop.
- Manage participants’ submissions of system description papers, manage
participants’ peer review of each others’ papers, and possibly shepherd
papers that need additional help in improving the writing.
- Review other task description papers.
- Define Roles for each Organizer:
- Lead Organizer - main point of contact, expected to ensure
deliverables are met on time and participate in contributing to
task duties
(see below).
- Co-Organizers - provide significant contributions to ensuring the
task runs smoothly. Some examples include, maintaining communication with
task participants, preparing data, creating and running
evaluation scripts,
and leading paper reviewing and acceptance.
- Advisory Organizers - more of a supervisor role, may not contribute
to detailed tasks but will provide guidance and support.
=== Important dates ===
- Task proposals due March 31, 2024 (Anywhere on Earth)
- Task selection notification May 18, 2024
=== Preliminary timetable ===
- Sample data ready July 15, 2024
- Training data ready September 1, 2024
- Evaluation data ready December 1, 2024 (internal deadline; not for public
release)
- Evaluation starts January 10, 2025
- Evaluation end by January 31, 2025 (latest date; task organizers may
choose an earlier date)
- Paper submission due February 2025
- Notification to authors on March 2025
- Camera-ready due April 2025
- SemEval workshop Summer 2025 (co-located with a major NLP conference)
Tasks that fail to keep up with crucial deadlines (such as the dates for
having the task and CodaLab website up and dates for uploading sample,
training, and evaluation data) or that diverge significantly from the
proposal may be cancelled at the discretion of SemEval organizers. While
consideration will be given to extenuating circumstances, our goal is to
provide sufficient time for the participants to develop strong and
well-thought-out systems. Cancelled tasks will be encouraged to submit
proposals for the subsequent year’s SemEval. To reduce the risk of tasks
failing to meet the deadlines, we are unlikely to accept multiple tasks
with overlap in the task organizers.
=== Submission Details ===
The task proposal should be a self-contained document of no longer than 3
pages (plus additional pages for references). All submissions must be in
PDF format, following the ACL template
<https://github.com/acl-org/acl-style-files>.
Each proposal should contain the following:
- Overview
- Summary of the task
- Why this task is needed and which communities would be interested
in participating
- Expected impact of the task
- Data & Resources
- How the training/testing data will be produced. Please discuss whether
existing corpora will be re-used.
- Details of copyright, so that the data can be used by the research
community both during the SemEval evaluation and afterwards
- How much data will be produced
- How data quality will be ensured and evaluated
- An example of what the data would look like
- Resources required to produce the data and prepare the task for
participants (annotation cost, annotation time, computation time, etc.)
- Assessment of any concerns with respect to ethics, privacy, or
security (e.g. personally identifiable information of private
individuals;
potential for systems to cause harm)
- Pilot Task (strongly recommended)
- Details of the pilot task
- What lessons were learned and how these will impact the task design
- Evaluation
- The evaluation methodology to be used, including clear evaluation
criteria
- For Task Reruns
- Justification for why a new iteration of the task is needed (see
criteria above)
- What will differ from the previous iteration
- Expected impact of the rerun compared with the previous iteration
- Task organizers
- Names, affiliations, email addresses
- (optional) brief description of relevant experience or expertise
- (if applicable) years and task numbers, of any SemEval tasks you
have run in the past
- Role of each organizer
Proposals will be reviewed by an independent group of area experts who may
not have familiarity with recent SemEval tasks, and therefore all proposals
should be written in a self-explanatory manner and contain sufficient
examples.
*The submission webpage is:* SemEval2025 Task Proposal Submission
<https://openreview.net/group?id=aclweb.org/ACL/2024/Workshop/SemEval> (
https://openreview.net/group?id=aclweb.org/ACL/2024/Workshop/SemEval)
For further information on this initiative, please refer to
https://semeval.github.io/SemEval2025/cft
=== Chairs ===
Atul Kr. Ojha, Insight SFI Centre for Data Analytics, DSI, University of
Galway
A. Seza Doğruöz, Ghent University
Giovanni Da San Martino, University of Padua
Harish Tayyar Madabushi, The University of Bath
Sara Rosenthal, IBM Research AI
Aiala Rosá, Universidad de la República - Uruguay
Contact: semevalorganizers(a)gmail.com