You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: _pages/home.md
+19-15Lines changed: 19 additions & 15 deletions
Original file line number
Diff line number
Diff line change
@@ -6,11 +6,16 @@ sitemap: false
6
6
permalink: /
7
7
---
8
8
9
-
We are a dynamic research group at the [Center for Information and Language Processing](https://www.cis.uni-muenchen.de/)located at [Ludwig Maximilian University Munich](https://www.en.uni-muenchen.de/index.html). Our primary focus is linguistically-informed Neural NLP: We use our deep understanding of language in our research and believe in the principle that learning is key to successful NLP -- the same way that the language capabilities of humans are based on learning.
9
+
We are a dynamic research group at the [Center for Information and Language Processing](https://www.cis.uni-muenchen.de/) at [Ludwig Maximilian University Munich](https://www.en.uni-muenchen.de/index.html), under the supervision of [Prof. Hinrich Schütze](https://www.cis.lmu.de/schuetze/). Our research areas include:
10
10
11
-
Our research areas are representation learning, multilinguality, machine learning for low-resource scenarios, cognitively motivated deep learning, linguistically informed deep learning (especially for morphology), digital humanities, and the intersection of NLP and robotics.
11
+
- Large Language Models (LLMs): We explore the behavior, structure, and potential of large-scale language models, examining their capabilities, biases, and self-assessment mechanisms to improve reliability and interpretability.
12
+
- Knowledge Expansion in NLP Models: We investigate how models can acquire and integrate new knowledge over time, using techniques that help improve their comprehension and generation abilities.
13
+
- Representation Learning: We study how language models represent linguistic and conceptual information by analyzing neurons and internal circuits to better understand and refine model behavior.
14
+
- Multilingual NLP: We address challenges in processing and evaluating multiple languages by developing benchmarks and methods for multilingual evaluation, including work on low-resource and non-standard languages.
15
+
- Intersection of NLP and Robotics: We integrate language understanding into robotic systems to enable natural, adaptable interaction in multimodal environments.
16
+
17
+
You can find resources (data, code, repositories) released by CIS lab members on our [GitHub page](https://github.com/cisnlp) and [HuggingFace page](https://huggingface.co/cis-lmu).
12
18
13
-
You can find resources (data, code, repositories) released by CIS lab members on our [GitHub page](https://github.com/cisnlp) and [HuggingFace page](https://huggingface.co/cis-lmu).
@@ -73,14 +78,13 @@ Our group publishes extensively in top NLP/ML conferences. Here is the summary o
73
78
| 2019 | 4 | 2 | 3 | - | 1 | 5 | 14 |
74
79
| 2018 | 4 | 3 | 1 | - | - | 5 | 13 |
75
80
-->
76
-
**We are looking for passionate new PhD students, Postdocs, interns, IDP (TUM) and MSc/BSc thesis students to join the team**[(more info)]({{ site.url }}{{ site.baseurl }}/vacancies) **!**
77
81
78
82
79
-
We are grateful for funding from the Munich Centre for Machine Learning ([MCML](https://mcml.ai/)), Elitenetzwerk Bayern (International Doctorate Program), Deutsche Forschungsgemeinschaft ([DFG](https://www.dfg.de/en/)) and from the European Research Council [ERC advanced grant](https://erc.europa.eu/funding/advanced-grants).
83
+
We are grateful for funding from the Munich Center for Machine Learning ([MCML](https://mcml.ai/)), the Elitenetzwerk Bayern (International Doctorate Program), the Deutsche Forschungsgemeinschaft ([DFG](https://www.dfg.de/en/)), and the European Research Council ([ERC Advanced Grant](https://erc.europa.eu/funding/advanced-grants)).
Copy file name to clipboardExpand all lines: _pages/openings.md
+1-3Lines changed: 1 addition & 3 deletions
Original file line number
Diff line number
Diff line change
@@ -8,9 +8,7 @@ permalink: /vacancies/
8
8
9
9
# Working at Schütze Lab
10
10
11
-
We are always looking for new group members with passion, talent, and grit!
12
-
13
-
You will have the chance to work at the cutting edge of research in computational linguistics and NLP. You may be exploring multilinguality on a Bible corpus of 1700+ languages, improving and enhancing latest transformer models or pushing the state of art on existing NLP tasks; you will be involved in asking the interesting questions, designing algorithms, conducting experiments and finally presenting your research in the best conferences around the world.
11
+
We are always looking for new group members with passion, talent, and grit! You will have the chance to work at the cutting edge of research in computational linguistics and NLP.
Copy file name to clipboardExpand all lines: _pages/projects.md
+12-15Lines changed: 12 additions & 15 deletions
Original file line number
Diff line number
Diff line change
@@ -6,30 +6,28 @@ sitemap: false
6
6
permalink: /projects/
7
7
---
8
8
9
-
# Funded Projects
9
+
# Projects
10
10
11
-
Our primary focus is linguistically-informed Neural Natural Language Processing (NNLP). Towards that end, we are working on various funded projects, as described below.
11
+
Our primary focus is linguistically-informed Neural Natural Language Processing. Towards that end, we are working on various funded projects, as described below.
12
12
13
-
## ERC Advanced Grant NonSequeToR
13
+
###ERC Advanced Grant NonSequeToR
14
14
15
15
Text understanding can fundamentally be viewed as a process of composition: the meaning of smaller units is composed to compute the meaning of larger units and eventually of sentences and documents. Our hypothesis is that optimal generalization in deep learning requires that more regular processes of composition are learned as composition functions whereas units that are the output of less regular processes are learned as static embeddings. We investigate novel representation learning algorithms and architectures to test this hypothesis. The envisioned goal of the project is a new robust and powerful text representation that
16
16
captures all aspects of form and meaning that NLP needs for successful processing of text.
17
+
See more details on [European Commission website](https://cordis.europa.eu/project/id/740516).
## High quality subword vocabulary induction (Project within the Munich Center for Machine Learning)
20
+
###High quality subword vocabulary induction (Project within the Munich Center for Machine Learning)
21
21
22
22
A common approach in representing text as input for deep learning models is to use heuristically induced word pieces. Such a representation is easier to process than characters, but does not incur the high cost of large word vocabularies. In this project, we will investigate alternatives to currently used heuristics that are a more natural representation of the semantics of text.
23
+
See more details on [MCML website](https://mcml.ai/areas_of_competence/).
## ReMLAV: Relational Machine Learning for Argument Validation (DFG project)
25
+
### ReMLAV: Relational Machine Learning for Argument Validation (DFG project)
27
26
28
27
Argument validation is the task of classifying a given argument as valid or invalid based on its linguistic form, the larger document context, world knowledge and other factors. This project aims to combine representation learning (both static and contextualized embeddings) and relational machine learning to solve this task.
28
+
See more details on [DFG website](https://gepris.dfg.de/gepris/projekt/376183703).
## Munich Doctoral Program (IDK): Premodern Cultures, Global Perspectives and the Foundations of a New Philology
30
+
### Munich Doctoral Program (IDK): Premodern Cultures, Global Perspectives and the Foundations of a New Philology
33
31
Based on an examination of 4000 years of history of literature, this program's goal is to synthesize theory and
34
32
practice of European traditions with those of the East Asian and South Asian cultural spheres as well as the Jewish and
35
33
Arab worlds. A particular focus will be on digital humanities methods.
@@ -38,11 +36,10 @@ Arab worlds. A particular focus will be on digital humanities methods.
38
36
39
37
<!-- ## ... and more. -->
40
38
41
-
# Robotics Projects
42
-
As large language models (LLMs) show impressive emergent abilities in various fields, we are exploring multiple directions combining LLMs into robotics.
43
-
Currently, we are mostly interested in the following topics:
As large language models (LLMs) show impressive emergent abilities in various fields, we are exploring multiple directions combining LLMs into robotics.
46
43
Given a high-level human instruction, the robot is supposed to understand the language instruction well and perform long-horizon memorizing and complex reasoning to complete the designated task.
47
44
We open-sourced a challenging benchmark *LoHoRavens* for this task and provided two baselines.
48
45
See more details on [LoHoRavens page](https://cisnlp.github.io/lohoravens-webpage/).
0 commit comments