Re "Wanderers, Kings, Merchants: The Story of India through Its Languages
by Peggy Mohan":
look forward to reading it. (History or historical events aside, does it
have to do with how language(s) became an indicator of
power/privilege/status? If humankind has anything in common, this could be
one general observation...)
It does to some extent, particularly with regard to the history of
languages and related things in India.
Re "If people are used to writing in their own language on computers, then
that language is more likely to survive":
I don't disagree, but ---
[Forewarning: one might not like my reply to the following, but I ask for
it to be interpreted with a scientific mindset rather, not with emotions
and sentiments related to language identity and particular cultural
practices.]
"Languages" (as in, particular language varieties, not "language" as in
language-at-large) come and go, born and die (and, sometimes, come back),
like trends, styles, cultures ("culture" as in a particular way of living,
a set of habits...). Esp. for users of varieties that have undergone
oppression/suppression, I understand that there is or can be much meaning
to many users in having the varieties be alive or in use. It is important
to the users. It is a symbol of their existence.
Yes, that is important.
But having witnessed how language has been abused (e.g. with research greed
by some CL/NLPers), I sometimes think one might have gone too far with how
much identity one attaches to any particular language.
Coming trom the background that I do, your statement above seems very
similar to saying that we might have gone too far with political
correctness or about opposing racism/misogyny/etc. Like most Europeans or
Americans, you seem to have no (or very little) idea about the toll that
discrimination -- even unintended discrimination -- takes on a very large
part of the human population.
In early 2009, I had written a rant in a blog post and the title of the
post was "English is Language Independent". I had made roughly the same
points which the now famous Bender Rule makes. I have been writing about
it, although that blog is now defunct. I did not, however, make any
proposal, as the Bender Rule does. I just pointed out the problem, so I am
in no way undermining the importance of that.
And one anonymous comment on this blog post was this: "Why don't you work
on a good project. I don't see the prejudice persisting for long once you
do that."
It is like saying about gender or racism: "Why don't you make
accomplishments equal to us? Once you do that, I don't see the prejudice
persisting for long."
Not to be picky here, but "I have heard some native speakers *users *of
some "Dravidian languages" say that there are some (I guess minor) problems
with Unicode for their languages": using the terms "native" and "speakers"
to refer to "users" (or as I sometimes use among knowers of language:
"languagers") has been an unhealthy baggage from our past practices in the
language space. I can't comment on the issue of "Dravidian languages" and
Unicode, but it seems one might need more info/details on the complaints to
act further.
Well, yes, the word "native" has indeed a very dark history. How could I
not know it? But, the manner in which linguists use the term "native
speaker" is very different. At least I think so. Note that I am only
mentioning linguists here, not "CL/NLPer". By the way, I thought I had
coined the term "NLPer" in my blog long ago, but I may be wrong about that.
I am pretty sure one can find out these details in some online forum or
some academic publication etc. I have not so far done that, but it is a
good idea to do that. I will try.
Re "psycholinguistic validity from computational validity":
in (cognitive/psycholinguistic) modeling, there is / can be not much
difference between the two. (When one has enough experiences with modeling
or with language/psycholinguistic phenomena, it's not hard to to see that
results from computational modeling could also hold elsewhere. The art then
is to be able to connect the two "realms". But then again, it depends on
the claims, of course.)
Yes, of course, it depends on the claim. The two realms can definitely be
connected. We don't disagree about that. In fact, I think, we don't really
disagree about that many things it seems. Even so, isn't it possible to
implement the same thing in many different ways when it comes to
computation? That is not the case with the brain/mind, of course. Here, I
am again making a distinction between computation and mathematics. Perhaps
you don't agree with that? In that case, perhaps we mean different things
by the term "computation".
Re science and engineering:
I am not sure if engineering has to be "just about" heuristics or short
cuts. There is good engineering and there is bad engineering, for the sake
of my arguments here. In the context of ML with language/textual data, one
ought to be careful with "computing based on values of surface
elements/strings".
Much of what the CL/NLP community/communities have been doing for the past
few decades has been "computing based on values of surface
elements/strings". This practice deserves serious re-evaluation (there are
lots of grey areas here and opportunities to compare processing across
finer granularities (without all the preprocessing
hacks/"heuristics"/"engineering") for various tasks and data types/formats,
without "words", "sentences", "linguistic structure(s)", "grammar" et al.).
I don't think of it as "it's engineering!", but some bad practices/culture
having been promoted as such and normalized (for a couple of decades?).
Good engineering can also be fine, thoughtful, and robust.
I completely agree. But sometimes I do work on things which, theoretically,
seem ridiculous to me, but they may be practically useful. At least half
(perhaps more) of my motive to work on language processing is to address
somehow, to any extent, the issue of linguistic empowerment. I am prepared
to compromise theoretically for that purpose.
Re "I don't think there is anything wrong with what you call grammar
hacking from the engineering point of view":
I do (think there is something wrong), because:
i. "all grammars leak" (from Sapir, also in Manning and Schütze (1999));
I know. I tell this to students every time I teach NLP or any related
subject.
ii. "words" (whatever they are) are too coarse-grained for computing.
I already agreed to that, but if they help in my benign motives, I am
prepared to use them.
Re "it [language(s)] is still likely to have an 'organic' structure":
couldn't that structure (one not associated with
"words"/"sentences"/"grammar") be one from math or computing? Or one that
is a by-product of a combination of these factors?
It certainly could.
Some CL/NLPers have made various claims concerning "structures" in the
past, borrowing the concepts from "linguistic structure(s)", from
"grammar". There was a lot of chiming along, many often have neglected the
fact that grammar could effect the impression of "structures" through
"words" etc. or that it all in turn patterns some of our thoughts/judgments
sub-/unconsciously. And the loop goes on.
(See also: See also:
https://twitter.com/adawan919/status/1532335891448057858)
Well, if you prefer the term "patterns" to "grammar" or "structure", I am
completely fine with that. As I said earlier, I am moving towards the
language games view of language, even for this exchange. We can't avoid
that if we are talking in a human/natural language. The only way to avoid
that, if there is a way, is to use only mathematical notation, but I don't
think we have reached that stage so far with the study of language.
Re "in English "John loves Mary" is in fact a very different thing than
"Mary loves John"":
one has to re-evaluate to which extent this matters in whichever form of
computing/computation one is engaged in and how often this "canonical form"
that you are implicitly referring to really occurs in data as well as how
this actually surfaces in data.
One should look at the data in front of one, not the framework/theory in
one's mind. (I believe in achieving better designs/systems through testing
from both a data-centric as well as an algorithm-centric perspective.
Hardware counts too!)
I mostly agree, but I am not sure whether you are saying that "John loves
Mary" may not perhaps be different from "Mary loves John"?
Re " it is unfair to blame Linguistics for that":
My focus in "[t]he "non-native speakers of X" has been a plague in
Linguistics" was on the "native" part. That has been my understanding, at
least to a great extent "nativeness" was so promoted/reinforced, esp.
within the school of generative Linguistics in the 2nd half of the 20th
century, when it comes to "linguistic judgments". I thought the propagation
stemmed from there. Who/What else do you think started it?
I think the word "native" was used in a derogatory/condescending way
throughout the English speaking world, even before the "birth of Modern
Linguistics". It was, in fact, the more polite word. One other common word
was "savage". I remember being shocked to find (very long ago) in Jane Eyre
the phrase "savages living on the banks of the Ganges". Savages? On the
bank of the Ganges?
But these usages are much older (than "Modern Linguistics").
The matter of "linguistic judgements" or "grammaticality" is very different
from that, regardless of what one's opinions are about the existence of
grammar.
All in all, your replies remind me of many of the reviewers' responses
"typical" of (i.e. I often got from) the *CL circle (of those who remained
in the past decade or so).
I don't know about that. I thought I had very unconventional views of NLP,
but I could be wrong about that, at least relatively.
If I may guess:
i. you don't have an academic background in Linguistics (esp. general
Linguistics,
That is true. I have learned about language(s) mostly on my own. So, if you
want me to show my degree in Linguistics, I have none, except a PhD in
Computational Linguistics. I was the second person to get a PhD
specifically in CL in India.
note that there is a difference between linguistics of particular languages
and that in a more general/theoretical manner (not about (p-)language
grammatical particularities),
Of course there is. What makes you think I don't know that? The fact is, my
knowledge of "(p)-language" is relatively very limited. Even my knowledge
of the syntax of Hindi (my "mother tongue"), in a formal sense, is very
limited. I mostly know about language in general.
ii. you learn about language(s) through mostly non-academic books or
through your own language experience(s) (which counts too, I am not
invalidating it/them here),
I have no idea what makes you say that. Am I supposed to list the
Linguistics books I have read, in addition to showing my Linguistics
degree? (Sorry if that sounds bitter, but it has happened in the past, not
literally, but effectively).
I can only say -- and it is strange to even have to say this -- that I
definitely know more about language in general and linguistic theory than
-- at least -- most graduates and postgraduates (including PhDs) of
Linguistics in India.
iii. you never had phonetics and phonology, nor
In my replies on this thread I have not mentioned anything related to
phonetics or phonology. So it must be from somewhere else that you have
this impression(?). Is it from some of the papers co-authored by me? I
think there are some which could give that impression. Explaining why that
could be so, will take this discussion somewhere else. I don't think it is
relevant here at all.
Is your point that I don't know about phonetics or phonology? If it is, I
would prefer not to answer that.
iv. do you realize how you can practice without "words"
I do. But I am also prepared to use "words" wherever they help. As I wrote
earlier, I have worked without "words" sometimes and have argued against
them.
--- did I get any right?
You got -- sort of -- the first and the third, assuming you were asking me
to show my Linguistics degree and whether I had formally taken courses on
Phonetics and Phonology.
I wanted to note this because --- and please do not take offense, it is not
meant personally for I respect your expertise and appreciate our exchanges
--- for a while, I didn't know where(else) to submit my findings. It wasn't
until I got all the rejections with rather shallow comments about language
(or language and computing) did I realize the "solidarity" one has built
with people with a background similar to yours might have been the driving
force of how some computational (general) linguists (as in, "general
language scientists" who also do computational work --- there are only a
few of us) got chased out of the arena. The "typical" excuses for practices
of this "culture" have been "engineering", "useful", "it works" --- but
without any/much grounding/interest in good generalizations. One puts
excess focus on processing but not on evaluation or interpretation. I think
it's time for a "culture" change in this regard.
To your other reply below (in triple quotes):
Sorry, but I didn't understand what you mean by triple quotes. I could find
any triple quotes in your comments.
re "language policy":
not everything has to be or can be regulated. Policies can help with
promoting/reinforcing/rectifying a particular situation/initiative.
I agree.
Forcing people e.g. to use language in one particular way or to use "one
language" only (whatever "language" "means"*)?
Again, I agree. However, like most Europeans and Americans, you seem to
have no (or very little) idea how people are already being forced to use
some language or another. And it is hardly a new phenomenon, but it has
become much more serious now due mainly to colonization and all its
effects. Do you have any idea how much hundreds of millions of Indians
suffer simply from being forced to use English? My primary motive in my
whole life, for good reasons, has been to counter linguistic
discrimination, mainly due to the imposition of English on any Indian who
has any ambition at all. I think that alone makes me sufficiently qualified
to "work on language". This is analogous to any other kind of
discrimination or prejudice.
I do not think that would be a good direction to go. For any regions, we
have seen both good/better and bad/worse policies throughout the course of
history. One would really have to evaluate the proposed policy in question
carefully.
I never said anything about forcing people to speak one language. That's
why I said I don't know what exactly the connection is with the language
policy. But it sure has a very strong connection, because the problem in
the first place is due to language policy, written and unwritten. Do you
know that there are and have been schools in the world, including India,
where students are punished if they are caught speaking in their mother
tongue (or first, "native" language). I still remember feeling recognition
and the impression it had on me when I read the famous novel about life in
Wales, How Green was My Valley. I had just become fluent in English then.
Depending on the situation, some may best change things through the
economy, some through government support, some through education and/or
grassroot-type of initiatives, some a combination of all these and more....
I agree.
*I have an answer to this... please wait for my next pub or so.
I would love to read it. I am eternally hungry for any fresh look on
language in general. Not so much for particular languages or varieties.
That is to me, to some extent, boring.
Re "it is very much like conservation of ecology or of species. I don't
think it (the latter) will be considered unwarranted prescriptivism":
see my 2nd response above.
The same for my reply to that comment.
Also, with language documentation, one can just document data without
promoting grammar. (That's probably the less unethical thing one can do
with language or language data.)
Again I agree. I never said anything about promoting grammar. I don't like
to read grammar books. It's painful to me, compared to almost any other
topic under the sun, except perhaps finance, commerce, and the intricacies
of legal procedures.
For the sake of completeness, I should clarify -- as it seems to matter --
that I have "never had Syntax or even Semantics or Pragmatics". I am mostly
self-taught, not just in Linguistics, but also in Computer Science and
almost everything else. Do you really think it matters in the context of
this discussion?
Again, for the sake of completeness, I should mention that for decades, I
have been reading all kinds of books that had anything to do with language,
mostly in general, but also about Hindi or Indian languages, not to mention
English. These have included what you call academic books on language in
general and about Linguistics. I still keep reading, as I know very well
that, being self-taught, I have some gaps in my knowledge of Linguistics
and Computer Science. My undergraduate degree was in Mechanical Engineering
(from 1990), but I hardly remember anything in that area. I have similarly
been reading all kinds of books for decades about computers and Computer
Science.
I am unable to see how any of this matters in the context of this
discussion.
By the way, I like the metaphor you use for language: It being like a
graphical user interface for the brain. That reminds me of the views of
Daniel Dennett about consciousness. He constantly compares elements making
up consciousness to graphical user interfaces on computers. Not that I
completely agree with him about consciousness, but I still find the
metaphor quite good, perhaps as an approximation.
Dear Sir/Ma'am,
I hope you are doing well and in good health. We are excited to announce a
call for a book chapter for an upcoming book titled "*Empowering
Low-Resource Languages With NLP Solutions.*"
Link: https://www.igi-global.com/publish/call-for-papers/call-details/6596
The objective of this book is to provide an in-depth understanding of
Natural Language Processing (NLP) techniques and applications specifically
tailored for low-resource languages. We believe that your valuable insights
and research in this domain would greatly enrich the content of this book.
To ensure a comprehensive and high-quality book, all submitted chapters
will undergo a rigorous peer-review process. The accepted book will be *indexed
in Scopus and Web of Science*, thereby enhancing the visibility and impact
of your work.
The book aims to cover a wide range of topics related to NLP in
low-resource languages. Some of the suggested topics, although not limited
to, include:
· Introduction to Low-Resource Languages in NLP
· Language Resource Acquisition for Low-Resource Languages
· Morphological Analysis and Morpho-Syntactic Processing
· Named Entity Recognition and Entity Linking for Low-Resource
Languages
· Part-of-Speech Tagging and Syntactic Parsing
· Machine Translation for Low-Resource Languages
· Sentiment Analysis and Opinion Mining for Low-Resource Languages
· Speech and Audio Processing for Low-Resource Languages
· Text Summarization and Information Retrieval for Low-Resource
Languages
· Multimodal NLP for Low-Resource Languages
· Code-switching and Language Identification for Low-Resource
Languages
· Evaluation and Benchmarking for NLP in Low-Resource Languages
· Applications of NLP in Low-Resource Language Settings
· Future Directions and Challenges in NLP
We encourage you to contribute a book chapter focusing on any of the
above-mentioned topics or related areas within the scope of NLP in
low-resource languages. The submission guidelines are as follows:
1. Please submit a chapter proposal (maximum 500 words) outlining the
objective, methodology, and expected outcomes of your proposed chapter by
August 15, 2023, to the submission portal:
https://www.igi-global.com/publish/call-for-papers/call-details/6596
2. Chapter proposals should include the title of the chapter, the
author(s) name, and their affiliations.
3. All submissions should be original and should not have been
previously published or currently under review elsewhere.
4. The chapters should be written in English and adhere to the
formatting guidelines provided after the acceptance of the proposal.
*Important Dates:*
August 15, 2023: Proposal Submission Deadline
August 25, 2023, 2023: Notification of Acceptance
September 17, 2023: Full Chapter Submission
October 31, 2023: Review Results Returned
December 12, 2023: Final Acceptance Notification
December 26, 2023: Final Chapter Submission
Thank you for considering this invitation, and we look forward to receiving
your valuable contribution to this book. If you have any further questions
or require additional information, please do not hesitate to contact us.
Best regards,
Editorial Team
Dr. Partha Pakray
National Institute of Technology Silchar
Email: partha(a)cse.nits.ac.in
Dr. Pankaj Dadure
University of Petroleum and Energy Studies Dehradun
Email: pankajk.dadure(a)ddn.upes.ac.in
Prof. Sivaji Bandyopadhyay
Jadavpur University, Kolkata
Email: sivaji.cse.ju(a)gmail.com
--
With Best Regards
Pankaj Dadure
Mobile: 9545757478
Third Call for papers
6th International Conference on Natural Language and Speech Processing
<http://icnlsp.org/>
We are delighted to invite you to ICNLSP 2023, which will be held virtually
from December 16th to 17th, 2023.
ICNLSP 2023 offers the opportunity for attendees (researchers, academics
and students, and industrials) to share their ideas and to connect to each
other and make them up to date on the ongoing research in the field.
ICNLSP 2023 aims to attract contributions related to natural language and
speech processing. Authors are invited to present their work relevant to
the topics of the conference.
The following list includes the topics of ICNLSP 2023 but not limited to:
Signal processing, acoustic modeling.
Architecture of speech recognition system.
Deep learning for speech recognition.
Analysis of speech.
Paralinguistics in Speech and Language.
Pathological speech and language.
Speech coding.
Speech comprehension.
Summarization.
Speech Translation.
Speech synthesis.
Speaker and language identification.
Phonetics, phonology and prosody.
Cognition and natural language processing.
Text categorization.
Sentiment analysis and opinion mining.
Computational Social Web.
Arabic dialects processing.
Under-resourced languages: tools and corpora.
New language models.
Arabic OCR.
Lexical semantics and knowledge representation.
Requirements engineering and NLP.
NLP tools for software requirements and engineering.
Knowledge fundamentals.
Knowledge management systems.
Information extraction.
Data mining and information retrieval.
Machine translation.
NLP for Arabic heritage documents.
*IMPORTANT DATES*
Submission deadline: *31 August 2023*
Notification of acceptance: *31 October 2023*
Camera-ready paper due: *20 November 2023*
Conference dates: *16, 17 December 2023*
*PUBLICATION*
1- All accepted papers will be published in ACL Anthology (
https://aclanthology.org/venues/icnlsp/).
2- Selected papers will be published in Signals and Communication
Technology (Springer) (https://www.springer.com/series/4748), indexed by
Scopus and zbMATH.
For more details, visit the conference website: https://www.icnlsp.org
*CONTACT*
icnlsp(at)gmail(dot)com
Best regards,
Mourad Abbas
*Call for Abstracts*
*'Towards Linguistically Motivated Computational Models of Framing'*
Date: Feb 28 - Mar 1, 2024
Location: Ruhr-University Bochum, Germany
Organizers: Annette Hautli-Janisz (University of Passau), Gabriella
Lapesa (University of Stuttgart), Ines Rehbein (University of Mannheim)
Homepage: https://sites.google.com/view/dgfs2024-framing
Call for Papers:
Framing is a central notion in the study of language use to rhetorically
package information strategically to achieve conversational goals
(Entman, 1993) but also, more broadly, in the study of how we organize
our experience (Goffman, 1974). In his seminal article, Entman (1993)
defines framing as "to select some aspects of a perceived reality and
make them more salient in a communicating text, in such a way as to
promote problem definition, causal interpretation, moral evaluation,
and/or treatment recommendation for the item described." This frame
definition has recently been operationalized in NLP in terms of
coarse-grained topic dimensions (Card et al., 2015), e.g., by modeling
the framing of immigration in the media as a challenge to economy vs. a
human rights issue. But there is more to frames than just topics.
The breadth of the debate on what constitutes a frame and on its (formal
and cognitive) definition naturally correlates to the interdisciplinary
relevance of this phenomenon: a theoretically motivated (computational)
model for framing is still needed, and this is precisely the goal of
this workshop, which will bring together researchers from theoretical,
applied and computational linguistics interested in framing analysis.
Our main interest is in furthering our understanding of how different
linguistic levels contribute to the framing of messages, and to pave the
way for the development of linguistically-driven computational models of
how people use framing to communicate their attitudes, preferences and
opinions.
We thus invite contributions that cover all levels of linguistic
analysis and methods: from phonetics (e.g., euphony: the use of
repetition, alliteration, rhymes and slogans to create persuasive
messages) and syntax (e.g., topicalization, passivization) to semantics
(lexical choices, such as Pro-Life vs. Pro-Choice; the use of pronouns
to create in- vs. out-groups; the use of metaphors; different types of
implicit meaning) to pragmatics (e.g., pragmatic framing through the use
of presupposition-triggering adverbs). We also invite work on
experimental and computational studies on framing which employ
linguistic structure to better understand instances of framing.
The workshop is part of the 46th Annual Conference of the German
Linguistic Society (DGfS 2024), held from 28 Feb - 1 March 2024 at
Ruhr-Universität Bochum, Germany.
*Submission instructions*:
We invite the submission of anonymous abstracts for 30 min talks
including discussion. Submissions should not exceed one page, 11pt
single spaced (abstract + references), with an optional additional page
for images. The reviewing process is double-blind; please ensure that
the paper does not include the authors' names and affiliations.
Furthermore, self-references that reveal the author's identity, e.g.,
"We previously showed (Smith, 1991) ...", should be avoided. Instead,
use citations such as "Smith previously showed (Smith, 1991) …".
*Submission deadline:* *August 25, 2023*
Abstract review period: Aug. 26, 2023 - Sept. 5, 2023
Meeting email: dgfs2024-framing(a)fim.uni-passau.de
--
Ines Rehbein
Data and Web Science Group
University of Mannheim, Germany
DLinNLP 2023 - Deep Learning Summer School at RANLP 2023
Second Call for Participation
Varna, Bulgaria
30th August - 1st September
https://dlinnlp2023.github.io/
We invite everyone interested in Machine Learning and Natural Language Processing to attend the Deep Learning Summer School at 14th biennial RANLP conference (RANLP 2023).
Purpose:
Deep Learning is a branch of machine learning that has gained significant traction in the field of Artificial Intelligence, pushing the envelope in the state-of-the-art, with many sub-areas including natural language, image, and speech processing employing it widely in their best-performing models.
This summer school will feature presentations from outstanding researchers in the field of Natural Language Processing (NLP) and Deep Learning. These will include coverage of recent advances in theoretical foundations and extensive practical coding sessions showcasing the latest relevant technology.
The summer school would be of interest to novices and established practitioners in the fields of NLP, corpus linguistics, language technologies, and similar related areas.
Important Dates:
30 August - 1 September: Deep Learning Summer School in NLP
Lectures:
* Lucas Beyer (Google Brain)
* Tharindu Ranasinghe (Aston University, UK)
* Iacer Calixto (University of Amsterdam, Holland)
Practical Sessions:
* Damith Premasiri (practical sessions) (University of Wolverhampton, UK)
* Isuri Anuradha (practical sessions) (University of Wolverhampton, UK)
* Anthony Hughes (practical sessions) (University of Wolverhampton, UK)
Registration:
**** Registration is now open: ******
https://ranlp.org/ranlp2023/index.php/fees-registration/
Programme:
Please refer to the website for the details of the programme:
https://dlinnlp2023.github.io/#programme
***There will be an informal poster presentation session where attendees can present their research work and get feedback from the experts in the field. ***
Contact Email: dlinnlp2023(a)gmail.com<mailto:dlinnlp2023@gmail.com>
[Note: the application deadline is Sunday 20 August]
The University of Gothenburg, Sweden, is offering four fully-funded PhD positions in computer science and engineering where the candidates can choose the project themselves out of fourteen options.
Two of the projects are related to NLP, one about efficient algorithms for corpus searching, and another about automatic generation of Wikipedia articles. See the ad for more information:
https://web103.reachmee.com/ext/I005/1035/job?site=7&lang=UK&validator=9b89…
The positions are fully funded for 5 years, including 20% teaching or other departemental duties.
Application deadline: 20 August 2023
best regards,
Peter Ljunglöf
------- ------ ----- ---- --- -- - - - - -
peter ljunglöf
peter.ljunglof(a)gu.se
data- och informationsteknik, och språkbanken
göteborgs universitet och chalmers tekniska högskola
-------------- --------- -------- ------- ------ ----- ---- --- -- - - - - -
Dear colleagues and friends,
The Research Center L3S invites applications for the position of a Research
Associate / PhD Candidate (m/f/d) “Computer Science: Knowledge Graphs &
Natural Language Processing” (Salary Scale 13 TV-L; 100 %) starting at the
earliest possible date. The position is limited to 3 years with the
possibility of extension. The regular weekly working time is 39.8 hours
(full-time).
*Description:*
The PhD topic will be in the context of the HybrInt project and the Open
Research Knowledge Graph (https://orkg.org/) focusing on building knowledge
graphs for the agricultural domain. The aim of these projects is to
research and develop NLP solutions as large language models (LLMs) for
crowdsourcing, representing and managing semantically structured, rich
representations of scholarly contributions and research data in knowledge
graphs and thus develop a novel model for scholarly communication. In the
context of the PhD thesis you will be responsible for building and
maintaining the ORKG data ingestion and processing pipelines to ensure the
flow of high-quality semantified resources from academic publications. Your
main responsibility in this position will be to build scalable solutions
that crawl, ingest, process publications, and thereby enrich the ORKG. You
will work alongside the ORKG engineering team to setup the AI/NLP ecosphere.
The tasks will focus on:
* Working in the areas of Natural Language Processing (text mining,
information extraction, information retrieval/search) and Machine Learning
of scholarly communication media (digital) data
* Research and development of suitable Large Language Models (LLMs) as NLP
solutions
* Conceptually designing, modeling, and implementing data-driven services
in one or more areas of information retrieval and extraction, data
enrichment, and linking of data
*Application Deadline:* 15.09.2023
*Web Address for Applications:* https://www.uni-hannover.de/en/jobs/id/6435/
(en); https://www.uni-hannover.de/de/jobs/id/6435/ (de)
Yours cordially,
Jennifer
Dear list members,
We are excited to announce that the AIGC corpus - The aiTECCL
Corpus, is now available to all members of the research community. The
aiTECCL Corpus was compiled by Jiajin Xu and Mingchen Sun of the Corpus
Research Group of the National Research Centre for Foreign Language
Education at Beijing Foreign Studies University.
The corpus consisting of two million words generated by the GPT-3.5
model, using identical writing prompts to those employed in *the TECCL
Corpus* <http://corpus.bfsu.edu.cn/info/1070/1449.htm>, aims to serve as a
reference corpus that exhibits a native-like linguistic quality. The corpus
is made available online on 9 August, 2023.
URL: http://114.251.154.212/cqp/
Username: test
Password: test
Please cite: Xu, Jiajin & Mingchen Sun. 2023. aiTECCL: An AIGC
English Essay Corpus. Beijing: National Research Centre for Foreign
Language Education, Beijing Foreign Studies University. Available online:
http://corpus.bfsu.edu.cn/info/1082/1913.htm
*Justifying the concept of "AIGC Corpus" (Artificial Intelligence Generated
Content Corpus) or Generative Corpus*
The creation of the AIGC Corpus helps expand the concept of
"corpus". In the classic definition of a corpus, the included materials
must be language samples that are authentically or naturally occurring in
real-life communication. Clearly, generative texts do not fall under this
category. We believe that the rationale for the generative corpus can be
viewed from at least three aspects:
1. The so-called principle of "authenticity" itself is a matter of
degree. For example, whether essays written by learners under exam
conditions belong to genuine communication is questionable. In existing
research, some elicited data also has authenticity issues similar to those
found in learners' interlanguage. Therefore, from the perspective of
existing corpora, there are texts with varying degrees of authenticity.
2. The generative corpus can serve as an essential complement to
existing corpora. The emergence of the generative corpus can reconcile the
distinction between "probable language" and "possible language." For
linguistic instances that have not yet appeared in reality, they can be
generated using large language models.
3. Creating a corpus using artificial intelligence technology is a
second-to-best solution under the current conditions for building specific
types of corpora. For example, the aiTECCL corpus simulates a reference
corpus of approximately 10,000 essays, close to the English native speaker
language quality, and written on the same topics as Chinese learners.
Without the use of artificial intelligence methods for generation, it might
be impossible to obtain a reference corpus of such quality and
comparability. Similarly, for corpus construction of languages from
least-developed countries or countries with extremely small populations,
without generative technology, it would be impossible to establish in the
short term.
Further details about the prompt and the Python script we utilised
to create the corpus will be provided on the site soon
<http://corpus.bfsu.edu.cn/info/1082/1913.htm>.
Best wishes,
Jiajin Xu
Ph.D., Professor
National Research Centre for Foreign Language Education
Beijing Foreign Studies University, China
Apologies for the multiple postings.
----
*Indian Language Summarization (ILSUM 2023)*
Website: https://ilsum.github.io/
To be organized in conjunction with FIRE 2023 (fire.irsi.res.in)
15th-18th December 2023, Goa, India
-------------------------------------------------------
The second shared task on Indian Language Summarization (ILSUM) aims at
creating an evaluation benchmark dataset for Indian Languages. This
year ILSUM consists of two subtasks
Subtask 1: This task builds upon the task from ILSUM 2022. In the
first edition, we covered two major Indian languages Hindi and
Gujarati alongside Indian English, a widely recognized dialect of the
English Language. This year's edition adds the Bengali language and an
expanded dataset for the languages from last year. Further, we will
provide abstractive summaries for a subset of each language (~1000 per
language) apart from the headlines which are semi-extractive summaries
in nature.
Like the previous edition, this will be a classic summarization task,
where we will provide
~15,000 article-summary pairs for each language and the participants are
expected to generate a fixed-length summary.
Subtask 2: The task is centred around identifying factual errors in
machine-generated summaries. With the recent implosion of Large
Language models, . While these LLMs are very good at summarization,
among other NLP tasks, they are often prone to hallucinations. This
means the model generates information that is not accurate, not based
on its training data, or is completely made up but looks accurate and
reliable. Further, such tools can be misused to generate misleading or
outright incorrect information. Identifying such inaccuracies can be a
challenging task.
Through this subtask, we aim to address the problem of identifying
factually incorrect information in LLM-generated summaries.
Participants will be provided with an article and its corresponding
machine-generated summary. The objective is to identify the presence
of factual incorrectness in the summaries if any, and classify them in
one of the predefined categories.
*Tentative Timeline*
-------------
7st August - Training Data Released and Registrations open
10th September - Test Data Release
20th September - Run Submission Deadline
25th September - Results Declared
10th October - Working notes due
25th October - Reviews Due
30th October - Camera Ready Submissions due
15th-18th December - FIRE 2023 at Goa, India
*Organisers*
----------------
Jagrat Patel, LDRP-ITR, Gandhinagar, India
Jaivin Barot, LDRP-ITR, Gandhinagar, India
Tanishka Gaur, LDRP-ITR, Gandhinagar, India
Shrey Satapara, Indian Institute of Technology, Hyderabad, India
Sandip Modha, LDRP-ITR, Gandhinagar, India
Parth Mehta, Parmonic, USA
Debasis Ganguly, University of Glasgow, Scotland
*For regular updates subscribe to our mailing list: **ilsum(a)googlegroups.com**
SECOND CALL FOR PAPERS
The 10th Workshop on Argument Mining @ EMNLP 2023
December 7, 2023
https://argmining-org.github.io/2023/
The 10th Workshop on Argument Mining will be held on December 7, 2023 in Singapore together with EMNLP 2023. This will be a hybrid event.
The Workshop on Argument Mining provides a regular forum for the presentation and discussion of cutting-edge research in argument mining (a.k.a argumentation mining) to both academic and industry researchers. By continuing a series of nine successful previous workshops, this edition will welcome the submission of long, short, and demo papers. It will feature two shared tasks, a panel on the last ten years of Argument Mining, and a keynote talk.
IMPORTANT DATES
- Direct paper submission deadline (Softconf): September 1, 2023
- Paper commitment from ARR: September 15, 2023
- Notification of acceptance: October 7, 2023
- Camera-ready submission: October 18, 2023
- Workshop: December 7, 2023
TOPICS OF INTEREST
The topics for submissions include but are not limited to:
- Automatic identification of argument components (e.g., premises and conclusions), the structure in which they form an argument, and relations between arguments and counterarguments (e.g., support and attack) in as well as across documents
- Automatic assessment of arguments and argumentation with respect to various properties, such as stance, clarity, and persuasiveness
- Automatic generation of arguments and their components, including the consideration of discourse goals (e.g., stages of a critical discussion or rhetorical strategies)
- Creation and evaluation of argument annotation schemes, relationships to linguistic and discourse annotations, (semi-) automatic argument annotation methods and tools, and creation of argumentation corpora
- Argument mining in specific genres and domains (e.g., social media, education, law, and scientific writing), each with a unique style (e.g., short informal text, highly structured writing, and long-form documents)
- Argument mining and generation from multi-modal and/or multilingual data
- Integration of commonsense and domain knowledge into argumentation models for mining and generation
- Combination of information retrieval methods with argument mining, e.g., in order to build the next generation of argumentative (web) search engines
- Real-world applications, including argument web search, opinion analysis in customer reviews, argument analysis in meetings, misinformation detection
- Perspectivist approaches to subjective argument mining tasks for which multiple ”ground truths” may exist, including multi-perspective machine learning and the creation of non-aggregated datasets
- Reflection on the ethical aspects and societal impact of argument mining methods
- Reflection on the future of argument mining in light of the fast advancement of large language models (LLMs).
SUBMISSIONS
The organizing committee welcomes the submission of long papers, short papers, and demo descriptions. Accepted papers will be presented either via oral or poster presentations. They will be included in the EMNLP proceedings as workshop papers.
- Long paper submissions must describe substantial, original, completed, and unpublished work. Wherever appropriate, concrete evaluation and analysis should be included. Long papers must be no longer than eight pages, including title, text, figures and tables. An unlimited number of pages is allowed for references. Two additional pages are allowed for appendices, and an extra page is allowed in the final version to address reviewers’ comments.
- Short paper submissions must describe original and unpublished work. Please note that a short paper is not a shortened long paper. Instead, short papers should have a point that can be made in a few pages, such as a small, focused contribution; a negative result; or an interesting application nugget. Short papers must be no longer than four pages, including title, text, figures and tables. An unlimited number of pages is allowed for references. One additional page is allowed for the appendix, and an extra page is allowed in the final version to address reviewers’ comments.
- Demo descriptions must be no longer than four pages, including title, text, examples, figures, tables, and references. A separate one-page document should be provided to the workshop organizers for demo descriptions, specifying furniture and equipment needed for the demo.
Multiple Submissions
ArgMining 2023 will not consider any paper that is under review in a journal or another conference or workshop at the time of submission, and submitted papers must not be submitted elsewhere during the review period.
ArgMining 2023 will also accept submissions of ARR-reviewed papers, provided that the ARR reviews and meta-reviews are available by the ARR commitment deadline (September 15). However, ArgMining 2023 will not accept direct submissions that are actively under review in ARR, or that overlap significantly (>25%) with such submissions.
Submission Format
All long, short, and demonstration submissions must follow the two-column EMNLP 2023 format. Authors are expected to use the LaTeX or Microsoft Word style template (https://2023.emnlp.org/calls/style-and-formatting/). Submissions must conform to the official EMNLP style guidelines, which are contained in these templates. Submissions must be electronic, in PDF format.
Submission Link and Deadline For Direct Submissions
Authors have to fill in the submission form in the START system and upload a PDF of their paper before September 1, 2023, 11:59 pm UTC-12h (anywhere on earth).
https://softconf.com/emnlp2023/ArgMining2023/
For the ARR commitment process, we will provide details in our second call for paper later in the summer.
Double Blind Review
ArgMining 2023 will follow the ACL policies for preserving the integrity of double-blind review for long and short paper submissions. Papers must not include authors’ names and affiliations. Furthermore, self-references or links (such as github) that reveal the author’s identity, e.g., “We previously showed (Smith, 1991) …” must be avoided. Instead, use citations such as “Smith previously showed (Smith, 1991) …” Papers that do not conform to these requirements will be rejected without review. Papers should not refer, for further detail, to documents that are not available to the reviewers. For example, do not omit or redact important citation information to preserve anonymity. Instead, use third person or named reference to this work, as described above (“Smith showed” rather than “we showed”). If important citations are not available to reviewers (e.g., awaiting publication), these paper/s should be anonymised and included in the appendix. They can then be referenced from the submission without compromising anonymity. Papers may be accompanied by a resource (software and/or data) described in the paper, but these resources should also be anonymized.
Unlike long and short papers, demo descriptions will not be anonymous. Demo descriptions should include the authors’ names and affiliations, and self-references are allowed.
ANONYMITY PERIOD (taken from the EMNLP call for papers in verbatim for the most part)
The following rules and guidelines are meant to protect the integrity of double-blind review and ensure that submissions are reviewed fairly. The rules make reference to the anonymity period, which runs from 1 month before the direct submission deadline (starting August 1, 2023) up to the date when your paper is accepted or rejected (October 7, 2023). For papers committed from ARR, the anonymity period starts August 15, 2023. Papers that are withdrawn during this period will no longer be subject to these rules.
You may not make a non-anonymized version of your paper available online to the general community (for example, via a preprint server) during the anonymity period. Versions of the paper include papers having essentially the same scientific content but possibly differing in minor details (including title and structure) and/or in length.
If you have posted a non-anonymized version of your paper online before the start of the anonymity period, you may submit an anonymized version to the conference. The submitted version must not refer to the non-anonymized version, and you must inform the programme chairs that a non-anonymized version exists.
You may not update the non-anonymized version during the anonymity period, and we ask you not to advertise it on social media or take other actions that would further compromise double-blind reviewing during the anonymity period.
You may make an anonymized version of your paper available (for example, on OpenReview), even during the anonymity period.
For arXiv submissions, August 1, 2023 11:59pm UTC-12h (anywhere on earth) is the latest time the paper can be uploaded if you plan a direct submission to the workshop (or August 15, 2023 for papers from ARR committed to the workshops on September 15, 2023).
BEST PAPER AWARDS
In order to recognize significant advancements in argument mining science and technology, ArgMining 2023 will include best paper awards. All papers at the workshop are eligible for the best paper awards and a selection committee consisting of prominent researchers in the fields of interest will select the recipients of the awards.
SHARED TASKS (Submission closed!)
We will be hosting two shared tasks this year:
- ImageArg-Shared-Task-2023: The First Shared Task in Multimodal Argument Mining
- PragTag-2023: The First Shared Task on Pragmatic Tagging of Peer Reviews
ArgMining 2023 ORGANIZING COMMITTEE
Milad Alshomary, Leibniz University Hannover, Germany
Chung-Chi Chen, National Institute of Advanced Industrial Science and Technology, Japan
Smaranda Muresan, Columbia University, USA
Joonsuk Park, University of Richmond, USA
Julia Romberg, Heinrich Heine University of Duesseldorf, Germany