## Fachgebiet Wissensverarbeitung (KDE), EECS, Universität Kassel

Das Fachgebiet Wissensverarbeitung des Fachbereichs Elektrotechnik/Informatik forscht an der Entwicklung von Methoden zur Wissensentdeckung und Wissensrepräsentation (Approximation und Exploration von Wissen, Ordnungsstrukturen in Wissen, Ontologieentwicklung) in Daten als auch in der Analyse von (sozialen) Netzwerkdaten und damit verbundenen Wissensprozessen (Metriken in Netzwerken, Anomalieerkennung, Charakterisierung von sozialen Netzwerken). Dabei liegt ein Schwerpunkt auf der exakten algebraischen Modellierung der verwendeten Strukturen und auf der Evaluierung und Neuentwicklung von Netzwerkmaßen. Neben der Erforschung von Grundlagen in den Gebieten Ordnungs- und Verbandstheorie, Beschreibungslogiken, Graphentheorie und Ontologie werden auch Anwendungen – bspw. in sozialen Medien sowie in der Szientometrie – erforscht.

Das Fachgebiet Wissensverarbeitung ist Mitglied im Wissenschaftlichen Zentrum für Informationstechnik-Gestaltung (ITeG) der Universität Kassel, im Wissenschaftlichen Zentrum INCHER der Universität Kassel und im Forschungszentrum L3S.

Testen Sie unser Social-Bookmark-System BibSonomy sowie unsere Namens-Suchmaschine Nameling!### Stellenangebot: Studentische Hilfskräfte – Causal Artificial Intelligence – bis 16.01.2023

### Unsere neusten Publikationen

- 1.Schäfermeier, B., Hirth, J., Hanika, T.: Research Topic Flows in Co-Authorship Networks. Scientometrics. (2022).In scientometrics, scientific collaboration is often analyzed by means of co-authorships. An aspect which is often overlooked and more difficult to quantify is the flow of expertise between authors from different research topics, which is an important part of scientific progress. With the Topic Flow Network (TFN) we propose a graph structure for the analysis of research topic flows between scientific authors and their respective research fields. Based on a multi-graph and a topic model, our proposed network structure accounts for intratopic as well as intertopic flows. Our method requires for the construction of a TFN solely a corpus of publications (i.e., author and abstract information). From this, research topics are discovered automatically through non-negative matrix factorization. The thereof derived TFN allows for the application of social network analysis techniques, such as common metrics and community detection. Most importantly, it allows for the analysis of intertopic flows on a large, macroscopic scale, i.e., between research topic, as well as on a microscopic scale, i.e., between certain sets of authors. We demonstrate the utility of TFNs by applying our method to two comprehensive corpora of altogether 20 Mio. publications spanning more than 60 years of research in the fields computer science and mathematics. Our results give evidence that TFNs are suitable, e.g., for the analysis of topical communities, the discovery of important authors in different fields, and, most notably, the analysis of intertopic flows, i.e., the transfer of topical expertise. Besides that, our method opens new directions for future research, such as the investigation of influence relationships between research fields.
@article{schafermeier2022research,

abstract = {In scientometrics, scientific collaboration is often analyzed by means of co-authorships. An aspect which is often overlooked and more difficult to quantify is the flow of expertise between authors from different research topics, which is an important part of scientific progress. With the Topic Flow Network (TFN) we propose a graph structure for the analysis of research topic flows between scientific authors and their respective research fields. Based on a multi-graph and a topic model, our proposed network structure accounts for intratopic as well as intertopic flows. Our method requires for the construction of a TFN solely a corpus of publications (i.e., author and abstract information). From this, research topics are discovered automatically through non-negative matrix factorization. The thereof derived TFN allows for the application of social network analysis techniques, such as common metrics and community detection. Most importantly, it allows for the analysis of intertopic flows on a large, macroscopic scale, i.e., between research topic, as well as on a microscopic scale, i.e., between certain sets of authors. We demonstrate the utility of TFNs by applying our method to two comprehensive corpora of altogether 20 Mio. publications spanning more than 60 years of research in the fields computer science and mathematics. Our results give evidence that TFNs are suitable, e.g., for the analysis of topical communities, the discovery of important authors in different fields, and, most notably, the analysis of intertopic flows, i.e., the transfer of topical expertise. Besides that, our method opens new directions for future research, such as the investigation of influence relationships between research fields.},

author = {Schäfermeier, Bastian and Hirth, Johannes and Hanika, Tom},

journal = {Scientometrics},

keywords = {co-authorships itegpub myown publist topic topic-flows topic-models},

month = 10,

title = {Research Topic Flows in Co-Authorship Networks},

year = 2022

}%0 Journal Article

%1 schafermeier2022research

%A Schäfermeier, Bastian

%A Hirth, Johannes

%A Hanika, Tom

%D 2022

%J Scientometrics

%R 10.1007/s11192-022-04529-w

%T Research Topic Flows in Co-Authorship Networks

%X In scientometrics, scientific collaboration is often analyzed by means of co-authorships. An aspect which is often overlooked and more difficult to quantify is the flow of expertise between authors from different research topics, which is an important part of scientific progress. With the Topic Flow Network (TFN) we propose a graph structure for the analysis of research topic flows between scientific authors and their respective research fields. Based on a multi-graph and a topic model, our proposed network structure accounts for intratopic as well as intertopic flows. Our method requires for the construction of a TFN solely a corpus of publications (i.e., author and abstract information). From this, research topics are discovered automatically through non-negative matrix factorization. The thereof derived TFN allows for the application of social network analysis techniques, such as common metrics and community detection. Most importantly, it allows for the analysis of intertopic flows on a large, macroscopic scale, i.e., between research topic, as well as on a microscopic scale, i.e., between certain sets of authors. We demonstrate the utility of TFNs by applying our method to two comprehensive corpora of altogether 20 Mio. publications spanning more than 60 years of research in the fields computer science and mathematics. Our results give evidence that TFNs are suitable, e.g., for the analysis of topical communities, the discovery of important authors in different fields, and, most notably, the analysis of intertopic flows, i.e., the transfer of topical expertise. Besides that, our method opens new directions for future research, such as the investigation of influence relationships between research fields. - 1.Stubbemann, M., Hanika, T., Schneider, F.M.: Intrinsic Dimension for Large-Scale Geometric Learning, https://arxiv.org/abs/2210.05301, (2022).
@misc{stubbemann2022intrinsic,

author = {Stubbemann, Maximilian and Hanika, Tom and Schneider, Friedrich Martin},

keywords = {2022 homepage kde kdepub loewe myown},

title = {Intrinsic Dimension for Large-Scale Geometric Learning},

year = 2022

}%0 Generic

%1 stubbemann2022intrinsic

%A Stubbemann, Maximilian

%A Hanika, Tom

%A Schneider, Friedrich Martin

%D 2022

%T Intrinsic Dimension for Large-Scale Geometric Learning

%U https://arxiv.org/abs/2210.05301 - 1.Hirth, J., Hanika, T.: Formal Conceptual Views in Neural Networks, http://arxiv.org/abs/2209.13517, (2022).Explaining neural network models is a challenging task that remains unsolved in its entirety to this day. This is especially true for high dimensional and complex data. With the present work, we introduce two notions for conceptual views of a neural network, specifically a many-valued and a symbolic view. Both provide novel analysis methods to enable a human AI analyst to grasp deeper insights into the knowledge that is captured by the neurons of a network. We test the conceptual expressivity of our novel views through different experiments on the ImageNet and Fruit-360 data sets. Furthermore, we show to which extent the views allow to quantify the conceptual similarity of different learning architectures. Finally, we demonstrate how conceptual views can be applied for abductive learning of human comprehensible rules from neurons. In summary, with our work, we contribute to the most relevant task of globally explaining neural networks models.
@misc{hirth2022formal,

abstract = {Explaining neural network models is a challenging task that remains unsolved in its entirety to this day. This is especially true for high dimensional and complex data. With the present work, we introduce two notions for conceptual views of a neural network, specifically a many-valued and a symbolic view. Both provide novel analysis methods to enable a human AI analyst to grasp deeper insights into the knowledge that is captured by the neurons of a network. We test the conceptual expressivity of our novel views through different experiments on the ImageNet and Fruit-360 data sets. Furthermore, we show to which extent the views allow to quantify the conceptual similarity of different learning architectures. Finally, we demonstrate how conceptual views can be applied for abductive learning of human comprehensible rules from neurons. In summary, with our work, we contribute to the most relevant task of globally explaining neural networks models.},

author = {Hirth, Johannes and Hanika, Tom},

keywords = {2022 NN conceptual fca kde kdepub myown network neural publist views},

note = {cite arxiv:2209.13517Comment: 17 pages, 8 figures, 9 tables},

title = {Formal Conceptual Views in Neural Networks},

year = 2022

}%0 Generic

%1 hirth2022formal

%A Hirth, Johannes

%A Hanika, Tom

%D 2022

%T Formal Conceptual Views in Neural Networks

%U http://arxiv.org/abs/2209.13517

%X Explaining neural network models is a challenging task that remains unsolved in its entirety to this day. This is especially true for high dimensional and complex data. With the present work, we introduce two notions for conceptual views of a neural network, specifically a many-valued and a symbolic view. Both provide novel analysis methods to enable a human AI analyst to grasp deeper insights into the knowledge that is captured by the neurons of a network. We test the conceptual expressivity of our novel views through different experiments on the ImageNet and Fruit-360 data sets. Furthermore, we show to which extent the views allow to quantify the conceptual similarity of different learning architectures. Finally, we demonstrate how conceptual views can be applied for abductive learning of human comprehensible rules from neurons. In summary, with our work, we contribute to the most relevant task of globally explaining neural networks models. - 1.Hanika, T., Hirth, J.: On the lattice of conceptual measurements. Information Sciences. 613, 453–468 (2022).We present a novel approach for data set scaling based on scale-measures from formal concept analysis, i.e., continuous maps between closure systems, for which we derive a canonical representation. Moreover, we prove that scale-measures can be lattice ordered using the canonical representation. This enables exploring the set of scale-measures by the use of meet and join operations. Furthermore we show that the lattice of scale-measures is isomorphic to the lattice of sub-closure systems that arises from the original data. Finally, we provide another representation of scale-measures using propositional logic in terms of data set features. Our theoretical findings are discussed by means of examples.
@article{HANIKA2022453,

abstract = {We present a novel approach for data set scaling based on scale-measures from formal concept analysis, i.e., continuous maps between closure systems, for which we derive a canonical representation. Moreover, we prove that scale-measures can be lattice ordered using the canonical representation. This enables exploring the set of scale-measures by the use of meet and join operations. Furthermore we show that the lattice of scale-measures is isomorphic to the lattice of sub-closure systems that arises from the original data. Finally, we provide another representation of scale-measures using propositional logic in terms of data set features. Our theoretical findings are discussed by means of examples.},

author = {Hanika, Tom and Hirth, Johannes},

journal = {Information Sciences},

keywords = {2022 FCA Lattice Measurements itegpub kde kdepub lattice measurement myown publist scale-measure scaling},

pages = {453-468},

title = {On the lattice of conceptual measurements},

volume = 613,

year = 2022

}%0 Journal Article

%1 HANIKA2022453

%A Hanika, Tom

%A Hirth, Johannes

%D 2022

%J Information Sciences

%P 453-468

%R https://doi.org/10.1016/j.ins.2022.09.005

%T On the lattice of conceptual measurements

%U https://www.sciencedirect.com/science/article/pii/S0020025522010489

%V 613

%X We present a novel approach for data set scaling based on scale-measures from formal concept analysis, i.e., continuous maps between closure systems, for which we derive a canonical representation. Moreover, we prove that scale-measures can be lattice ordered using the canonical representation. This enables exploring the set of scale-measures by the use of meet and join operations. Furthermore we show that the lattice of scale-measures is isomorphic to the lattice of sub-closure systems that arises from the original data. Finally, we provide another representation of scale-measures using propositional logic in terms of data set features. Our theoretical findings are discussed by means of examples. - 1.Dürrschnabel, D., Hanika, T., Stubbemann, M.: FCA2VEC: Embedding Techniques for Formal Concept Analysis. In: Missaoui, R., Kwuida, L., and Abdessalem, T. (eds.) Complex Data Analytics with Formal Concept Analysis. pp. 47–74. Springer International Publishing (2022).
@incollection{DBLP:books/sp/missaoui2022/DurrschnabelHS22,

author = {Dürrschnabel, Dominik and Hanika, Tom and Stubbemann, Maximilian},

booktitle = {Complex Data Analytics with Formal Concept Analysis},

editor = {Missaoui, Rokia and Kwuida, Léonard and Abdessalem, Talel},

keywords = {2022 embedding fca itegpub kdepub machinelearning myown publist},

pages = {47--74},

publisher = {Springer International Publishing},

title = {FCA2VEC: Embedding Techniques for Formal Concept Analysis},

year = 2022

}%0 Book Section

%1 DBLP:books/sp/missaoui2022/DurrschnabelHS22

%A Dürrschnabel, Dominik

%A Hanika, Tom

%A Stubbemann, Maximilian

%B Complex Data Analytics with Formal Concept Analysis

%D 2022

%E Missaoui, Rokia

%E Kwuida, Léonard

%E Abdessalem, Talel

%I Springer International Publishing

%P 47--74

%R 10.1007/978-3-030-93278-7_3

%T FCA2VEC: Embedding Techniques for Formal Concept Analysis

%U https://doi.org/10.1007/978-3-030-93278-7_3 - 1.Felde, M., Koyda, M.: Interval-Dismantling for Lattices, https://arxiv.org/abs/2208.01479, (2022).Dismantling allows for the removal of elements of a set, or in our case lattice, without disturbing the remaining structure. In this paper we have extended the notion of dismantling by single elements to the dismantling by intervals in a lattice. We utilize theory from Formal Concept Analysis (FCA) to show that lattices dismantled by intervals correspond to closed subrelations in the respective formal context, and that there exists a unique kernel with respect to dismantling by intervals. Furthermore, we show that dismantling intervals can be identified directly in the formal context utilizing a characterization via arrow relations and provide an algorithm to compute all dismantling intervals.
@preprint{felde2022intervaldismantling,

abstract = {Dismantling allows for the removal of elements of a set, or in our case lattice, without disturbing the remaining structure. In this paper we have extended the notion of dismantling by single elements to the dismantling by intervals in a lattice. We utilize theory from Formal Concept Analysis (FCA) to show that lattices dismantled by intervals correspond to closed subrelations in the respective formal context, and that there exists a unique kernel with respect to dismantling by intervals. Furthermore, we show that dismantling intervals can be identified directly in the formal context utilizing a characterization via arrow relations and provide an algorithm to compute all dismantling intervals.},

author = {Felde, Maximilian and Koyda, Maren},

keywords = {arrow-relations concepts context dismantling fca myown},

note = {cite arxiv:2208.01479Comment: 12 pages, 5 figures, 1 algorithm},

title = {Interval-Dismantling for Lattices},

year = 2022

}%0 Generic

%1 felde2022intervaldismantling

%A Felde, Maximilian

%A Koyda, Maren

%D 2022

%R 10.48550/arXiv.2208.01479

%T Interval-Dismantling for Lattices

%U https://arxiv.org/abs/2208.01479

%X Dismantling allows for the removal of elements of a set, or in our case lattice, without disturbing the remaining structure. In this paper we have extended the notion of dismantling by single elements to the dismantling by intervals in a lattice. We utilize theory from Formal Concept Analysis (FCA) to show that lattices dismantled by intervals correspond to closed subrelations in the respective formal context, and that there exists a unique kernel with respect to dismantling by intervals. Furthermore, we show that dismantling intervals can be identified directly in the formal context utilizing a characterization via arrow relations and provide an algorithm to compute all dismantling intervals. - 1.Felde, M., Stumme, G.: Attribute Exploration with Multiple Contradicting Partial Experts. In: Braun, T., Cristea, D., and Jäschke, R. (eds.) Graph-Based Representation and Reasoning. pp. 51–65. Springer International Publishing, Cham (2022).Attribute exploration is a method from Formal Concept Analysis (FCA) that helps a domain expert discover structural dependencies in knowledge domains which can be represented as formal contexts (cross tables of objects and attributes). In this paper we present an extension of attribute exploration that allows for a group of domain experts and explores their shared views. Each expert has their own view of the domain and the views of multiple experts may contain contradicting information.
@inproceedings{10.1007/978-3-031-16663-1_5,

abstract = {Attribute exploration is a method from Formal Concept Analysis (FCA) that helps a domain expert discover structural dependencies in knowledge domains which can be represented as formal contexts (cross tables of objects and attributes). In this paper we present an extension of attribute exploration that allows for a group of domain experts and explores their shared views. Each expert has their own view of the domain and the views of multiple experts may contain contradicting information.},

address = {Cham},

author = {Felde, Maximilian and Stumme, Gerd},

booktitle = {Graph-Based Representation and Reasoning},

editor = {Braun, Tanya and Cristea, Diana and Jäschke, Robert},

keywords = {attribute-exploration fca multiple-experts myown},

pages = {51--65},

publisher = {Springer International Publishing},

title = {Attribute Exploration with Multiple Contradicting Partial Experts},

year = 2022

}%0 Conference Paper

%1 10.1007/978-3-031-16663-1_5

%A Felde, Maximilian

%A Stumme, Gerd

%B Graph-Based Representation and Reasoning

%C Cham

%D 2022

%E Braun, Tanya

%E Cristea, Diana

%E Jäschke, Robert

%I Springer International Publishing

%P 51--65

%R 10.1007/978-3-031-16663-1_5

%T Attribute Exploration with Multiple Contradicting Partial Experts

%X Attribute exploration is a method from Formal Concept Analysis (FCA) that helps a domain expert discover structural dependencies in knowledge domains which can be represented as formal contexts (cross tables of objects and attributes). In this paper we present an extension of attribute exploration that allows for a group of domain experts and explores their shared views. Each expert has their own view of the domain and the views of multiple experts may contain contradicting information.

%@ 978-3-031-16663-1 - 1.Schäfermeier, B., Stumme, G., Hanika, T.: Mapping Research Trajectories, https://arxiv.org/abs/2204.11859, (2022).
@misc{https://doi.org/10.48550/arxiv.2204.11859,

author = {Schäfermeier, Bastian and Stumme, Gerd and Hanika, Tom},

keywords = {mapping myown publist research trajectories},

publisher = {arXiv},

title = {Mapping Research Trajectories},

year = 2022

}%0 Generic

%1 https://doi.org/10.48550/arxiv.2204.11859

%A Schäfermeier, Bastian

%A Stumme, Gerd

%A Hanika, Tom

%D 2022

%I arXiv

%R 10.48550/ARXIV.2204.11859

%T Mapping Research Trajectories

%U https://arxiv.org/abs/2204.11859 - 1.Hanika, T., Hirth, J.: Knowledge cores in large formal contexts. Annals of Mathematics and Artificial Intelligence. (2022).Knowledge computation tasks, such as computing a base of valid implications, are often infeasible for large data sets. This is in particular true when deriving canonical bases in formal concept analysis (FCA). Therefore, it is necessary to find techniques that on the one hand reduce the data set size, but on the other hand preserve enough structure to extract useful knowledge. Many successful methods are based on random processes to reduce the size of the investigated data set. This, however, makes them hardly interpretable with respect to the discovered knowledge. Other approaches restrict themselves to highly supported subsets and omit rare and (maybe) interesting patterns. An essentially different approach is used in network science, called k-cores. These cores are able to reflect rare patterns, as long as they are well connected within the data set. In this work, we study k-cores in the realm of FCA by exploiting the natural correspondence of bi-partite graphs and formal contexts. This structurally motivated approach leads to a comprehensible extraction of knowledge cores from large formal contexts.
@article{Hanika2022,

abstract = {Knowledge computation tasks, such as computing a base of valid implications, are often infeasible for large data sets. This is in particular true when deriving canonical bases in formal concept analysis (FCA). Therefore, it is necessary to find techniques that on the one hand reduce the data set size, but on the other hand preserve enough structure to extract useful knowledge. Many successful methods are based on random processes to reduce the size of the investigated data set. This, however, makes them hardly interpretable with respect to the discovered knowledge. Other approaches restrict themselves to highly supported subsets and omit rare and (maybe) interesting patterns. An essentially different approach is used in network science, called k-cores. These cores are able to reflect rare patterns, as long as they are well connected within the data set. In this work, we study k-cores in the realm of FCA by exploiting the natural correspondence of bi-partite graphs and formal contexts. This structurally motivated approach leads to a comprehensible extraction of knowledge cores from large formal contexts.},

author = {Hanika, Tom and Hirth, Johannes},

journal = {Annals of Mathematics and Artificial Intelligence},

keywords = {2022 bigdata bipartite cores fca itegpub k-cores kde kdepub myown publist sai},

month = {apr},

title = {Knowledge cores in large formal contexts},

year = 2022

}%0 Journal Article

%1 Hanika2022

%A Hanika, Tom

%A Hirth, Johannes

%D 2022

%J Annals of Mathematics and Artificial Intelligence

%R 10.1007/s10472-022-09790-6

%T Knowledge cores in large formal contexts

%U https://doi.org/10.1007/s10472-022-09790-6

%X Knowledge computation tasks, such as computing a base of valid implications, are often infeasible for large data sets. This is in particular true when deriving canonical bases in formal concept analysis (FCA). Therefore, it is necessary to find techniques that on the one hand reduce the data set size, but on the other hand preserve enough structure to extract useful knowledge. Many successful methods are based on random processes to reduce the size of the investigated data set. This, however, makes them hardly interpretable with respect to the discovered knowledge. Other approaches restrict themselves to highly supported subsets and omit rare and (maybe) interesting patterns. An essentially different approach is used in network science, called k-cores. These cores are able to reflect rare patterns, as long as they are well connected within the data set. In this work, we study k-cores in the realm of FCA by exploiting the natural correspondence of bi-partite graphs and formal contexts. This structurally motivated approach leads to a comprehensible extraction of knowledge cores from large formal contexts. - 1.Stubbemann, M., Stumme, G.: LG4AV: Combining Language Models and Graph Neural Networks for Author Verification. In: Bouadi, T., Fromont, E., and Hüllermeier, E. (eds.) Advances in Intelligent Data Analysis XX. pp. 315–326. Springer International Publishing, Cham (2022).The verification of document authorships is important in various settings. Researchers are for example judged and compared by the amount and impact of their publications and public figures are confronted by their posts on social media. Therefore, it is important that authorship information in frequently used data sets is correct. The question whether a given document is written by a given author is commonly referred to as authorship verification (AV). While AV is a widely investigated problem in general, only few works consider settings where the documents are short and written in a rather uniform style. This makes most approaches impractical for bibliometric data. Here, authorships of scientific publications have to be verified, often with just abstracts and titles available. To this point, we present LG4AV which combines language models and graph neural networks for authorship verification. By directly feeding the available texts in a pre-trained transformer architecture, our model does not need any hand-crafted stylometric features that are not meaningful in scenarios where the writing style is, at least to some extent, standardized. By the incorporation of a graph neural network structure, our model can benefit from relations between authors that are meaningful with respect to the verification process.
@inproceedings{10.1007/978-3-031-01333-1_25,

abstract = {The verification of document authorships is important in various settings. Researchers are for example judged and compared by the amount and impact of their publications and public figures are confronted by their posts on social media. Therefore, it is important that authorship information in frequently used data sets is correct. The question whether a given document is written by a given author is commonly referred to as authorship verification (AV). While AV is a widely investigated problem in general, only few works consider settings where the documents are short and written in a rather uniform style. This makes most approaches impractical for bibliometric data. Here, authorships of scientific publications have to be verified, often with just abstracts and titles available. To this point, we present LG4AV which combines language models and graph neural networks for authorship verification. By directly feeding the available texts in a pre-trained transformer architecture, our model does not need any hand-crafted stylometric features that are not meaningful in scenarios where the writing style is, at least to some extent, standardized. By the incorporation of a graph neural network structure, our model can benefit from relations between authors that are meaningful with respect to the verification process.},

address = {Cham},

author = {Stubbemann, Maximilian and Stumme, Gerd},

booktitle = {Advances in Intelligent Data Analysis XX},

editor = {Bouadi, Tassadit and Fromont, Elisa and Hüllermeier, Eyke},

keywords = {2022 homepage kdepub myown regio},

pages = {315--326},

publisher = {Springer International Publishing},

title = {LG4AV: Combining Language Models and Graph Neural Networks for Author Verification},

year = 2022

}%0 Conference Paper

%1 10.1007/978-3-031-01333-1_25

%A Stubbemann, Maximilian

%A Stumme, Gerd

%B Advances in Intelligent Data Analysis XX

%C Cham

%D 2022

%E Bouadi, Tassadit

%E Fromont, Elisa

%E Hüllermeier, Eyke

%I Springer International Publishing

%P 315--326

%T LG4AV: Combining Language Models and Graph Neural Networks for Author Verification

%U https://link.springer.com/chapter/10.1007/978-3-031-01333-1_25

%X The verification of document authorships is important in various settings. Researchers are for example judged and compared by the amount and impact of their publications and public figures are confronted by their posts on social media. Therefore, it is important that authorship information in frequently used data sets is correct. The question whether a given document is written by a given author is commonly referred to as authorship verification (AV). While AV is a widely investigated problem in general, only few works consider settings where the documents are short and written in a rather uniform style. This makes most approaches impractical for bibliometric data. Here, authorships of scientific publications have to be verified, often with just abstracts and titles available. To this point, we present LG4AV which combines language models and graph neural networks for authorship verification. By directly feeding the available texts in a pre-trained transformer architecture, our model does not need any hand-crafted stylometric features that are not meaningful in scenarios where the writing style is, at least to some extent, standardized. By the incorporation of a graph neural network structure, our model can benefit from relations between authors that are meaningful with respect to the verification process.

%@ 978-3-031-01333-1 - 1.Hanika, T., Schneider, F.M., Stumme, G.: Intrinsic dimension of geometric data sets. Tohoku Mathematical Journal. 74, 23–52 (2022).The curse of dimensionality is a phenomenon frequently observed in machine learning (ML) and knowledge discovery (KD). There is a large body of literature investigating its origin and impact, using methods from mathematics as well as from computer science. Among the mathematical insights into data dimensionality, there is an intimate link between the dimension curse and the phenomenon of measure concentration, which makes the former accessible to methods of geometric analysis. The present work provides a comprehensive study of the intrinsic geometry of a data set, based on Gromov's metric measure geometry and Pestov's axiomatic approach to intrinsic dimension. In detail, we define a concept of geometric data set and introduce a metric as well as a partial order on the set of isomorphism classes of such data sets. Based on these objects, we propose and investigate an axiomatic approach to the intrinsic dimension of geometric data sets and establish a concrete dimension function with the desired properties. Our model for data sets and their intrinsic dimension is computationally feasible and, moreover, adaptable to specific ML/KD-algorithms, as illustrated by various experiments.
@article{10.2748/tmj.20201015a,

abstract = {The curse of dimensionality is a phenomenon frequently observed in machine learning (ML) and knowledge discovery (KD). There is a large body of literature investigating its origin and impact, using methods from mathematics as well as from computer science. Among the mathematical insights into data dimensionality, there is an intimate link between the dimension curse and the phenomenon of measure concentration, which makes the former accessible to methods of geometric analysis. The present work provides a comprehensive study of the intrinsic geometry of a data set, based on Gromov's metric measure geometry and Pestov's axiomatic approach to intrinsic dimension. In detail, we define a concept of geometric data set and introduce a metric as well as a partial order on the set of isomorphism classes of such data sets. Based on these objects, we propose and investigate an axiomatic approach to the intrinsic dimension of geometric data sets and establish a concrete dimension function with the desired properties. Our model for data sets and their intrinsic dimension is computationally feasible and, moreover, adaptable to specific ML/KD-algorithms, as illustrated by various experiments.},

author = {Hanika, Tom and Schneider, Friedrich Martin and Stumme, Gerd},

journal = {Tohoku Mathematical Journal},

keywords = {dcd fca intrinsic itegpub kde kdepub loewe myown publist},

number = 1,

pages = {23 -- 52},

publisher = {Tohoku University, Mathematical Institute},

title = {Intrinsic dimension of geometric data sets},

volume = 74,

year = 2022

}%0 Journal Article

%1 10.2748/tmj.20201015a

%A Hanika, Tom

%A Schneider, Friedrich Martin

%A Stumme, Gerd

%D 2022

%I Tohoku University, Mathematical Institute

%J Tohoku Mathematical Journal

%N 1

%P 23 -- 52

%R 10.2748/tmj.20201015a

%T Intrinsic dimension of geometric data sets

%U https://doi.org/10.2748/tmj.20201015a

%V 74

%X The curse of dimensionality is a phenomenon frequently observed in machine learning (ML) and knowledge discovery (KD). There is a large body of literature investigating its origin and impact, using methods from mathematics as well as from computer science. Among the mathematical insights into data dimensionality, there is an intimate link between the dimension curse and the phenomenon of measure concentration, which makes the former accessible to methods of geometric analysis. The present work provides a comprehensive study of the intrinsic geometry of a data set, based on Gromov's metric measure geometry and Pestov's axiomatic approach to intrinsic dimension. In detail, we define a concept of geometric data set and introduce a metric as well as a partial order on the set of isomorphism classes of such data sets. Based on these objects, we propose and investigate an axiomatic approach to the intrinsic dimension of geometric data sets and establish a concrete dimension function with the desired properties. Our model for data sets and their intrinsic dimension is computationally feasible and, moreover, adaptable to specific ML/KD-algorithms, as illustrated by various experiments. - 1.Dürrschnabel, D., Stumme, G.: Force-Directed Layout of Order Diagrams Using Dimensional Reduction. In: Braud, A., Buzmakov, A., Hanika, T., and Le Ber, F. (eds.) Formal Concept Analysis. pp. 224–240. Springer International Publishing, Cham (2021).Order diagrams allow human analysts to understand and analyze structural properties of ordered data. While an expert can create easily readable order diagrams, the automatic generation of those remains a hard task. In this work, we adapt force-directed approaches, which are known to generate aesthetically-pleasing drawings of graphs, to the realm of order diagrams. Our algorithm ReDraw thereby embeds the order in a high dimension and then iteratively reduces the dimension until a two-dimensional drawing is achieved. To improve aesthetics, this reduction is equipped with two force-directed steps where one step optimizes the distances of nodes and the other one the distances of lines in order to satisfy a set of a priori fixed conditions. By respecting an invariant about the vertical position of the elements in each step of our algorithm we ensure that the resulting drawings satisfy all necessary properties of order diagrams. Finally, we present the results of a user study to demonstrate that our algorithm outperforms comparable approaches on drawings of lattices with a high degree of distributivity.
@inproceedings{10.1007/978-3-030-77867-5_14,

abstract = {Order diagrams allow human analysts to understand and analyze structural properties of ordered data. While an expert can create easily readable order diagrams, the automatic generation of those remains a hard task. In this work, we adapt force-directed approaches, which are known to generate aesthetically-pleasing drawings of graphs, to the realm of order diagrams. Our algorithm ReDraw thereby embeds the order in a high dimension and then iteratively reduces the dimension until a two-dimensional drawing is achieved. To improve aesthetics, this reduction is equipped with two force-directed steps where one step optimizes the distances of nodes and the other one the distances of lines in order to satisfy a set of a priori fixed conditions. By respecting an invariant about the vertical position of the elements in each step of our algorithm we ensure that the resulting drawings satisfy all necessary properties of order diagrams. Finally, we present the results of a user study to demonstrate that our algorithm outperforms comparable approaches on drawings of lattices with a high degree of distributivity.},

address = {Cham},

author = {Dürrschnabel, Dominik and Stumme, Gerd},

booktitle = {Formal Concept Analysis},

editor = {Braud, Agnès and Buzmakov, Aleksey and Hanika, Tom and Le Ber, Florence},

keywords = {2021 diagram_drawing itegpub lattices myown order_diagrams spring_embedder},

pages = {224--240},

publisher = {Springer International Publishing},

title = {Force-Directed Layout of Order Diagrams Using Dimensional Reduction},

year = 2021

}%0 Conference Paper

%1 10.1007/978-3-030-77867-5_14

%A Dürrschnabel, Dominik

%A Stumme, Gerd

%B Formal Concept Analysis

%C Cham

%D 2021

%E Braud, Agnès

%E Buzmakov, Aleksey

%E Hanika, Tom

%E Le Ber, Florence

%I Springer International Publishing

%P 224--240

%T Force-Directed Layout of Order Diagrams Using Dimensional Reduction

%X Order diagrams allow human analysts to understand and analyze structural properties of ordered data. While an expert can create easily readable order diagrams, the automatic generation of those remains a hard task. In this work, we adapt force-directed approaches, which are known to generate aesthetically-pleasing drawings of graphs, to the realm of order diagrams. Our algorithm ReDraw thereby embeds the order in a high dimension and then iteratively reduces the dimension until a two-dimensional drawing is achieved. To improve aesthetics, this reduction is equipped with two force-directed steps where one step optimizes the distances of nodes and the other one the distances of lines in order to satisfy a set of a priori fixed conditions. By respecting an invariant about the vertical position of the elements in each step of our algorithm we ensure that the resulting drawings satisfy all necessary properties of order diagrams. Finally, we present the results of a user study to demonstrate that our algorithm outperforms comparable approaches on drawings of lattices with a high degree of distributivity.

%@ 978-3-030-77867-5 - 1.Hanika, T., Hirth, J.: Exploring Scale-Measures of Data Sets. In: Braud, A., Buzmakov, A., Hanika, T., and Ber, F.L. (eds.) Formal Concept Analysis - 16th International Conference, ICFCA 2021, Strasbourg, France, June 29 - July 2, 2021, Proceedings. pp. 261–269. Springer (2021).
@inproceedings{DBLP:conf/icfca/HanikaH21,

author = {Hanika, Tom and Hirth, Johannes},

booktitle = {Formal Concept Analysis - 16th International Conference, ICFCA 2021, Strasbourg, France, June 29 - July 2, 2021, Proceedings},

editor = {Braud, Agnès and Buzmakov, Aleksey and Hanika, Tom and Ber, Florence Le},

keywords = {closure closure-system exploring fca lattice myown sai scaling},

pages = {261--269},

publisher = {Springer},

series = {Lecture Notes in Computer Science},

title = {Exploring Scale-Measures of Data Sets},

volume = 12733,

year = 2021

}%0 Conference Paper

%1 DBLP:conf/icfca/HanikaH21

%A Hanika, Tom

%A Hirth, Johannes

%B Formal Concept Analysis - 16th International Conference, ICFCA 2021, Strasbourg, France, June 29 - July 2, 2021, Proceedings

%D 2021

%E Braud, Agnès

%E Buzmakov, Aleksey

%E Hanika, Tom

%E Ber, Florence Le

%I Springer

%P 261--269

%R 10.1007/978-3-030-77867-5_17

%T Exploring Scale-Measures of Data Sets

%U https://doi.org/10.1007/978-3-030-77867-5_17

%V 12733 - 1.Hanika, T., Hirth, J.: Quantifying the Conceptual Error in Dimensionality Reduction. In: Braun, T., Gehrke, M., Hanika, T., and Hernandez, N. (eds.) Graph-Based Representation and Reasoning - 26th International Conference on Conceptual Structures, ICCS 2021, Virtual Event, September 20-22, 2021, Proceedings. pp. 105–118. Springer (2021).
@inproceedings{DBLP:conf/iccs/HanikaH21,

author = {Hanika, Tom and Hirth, Johannes},

booktitle = {Graph-Based Representation and Reasoning - 26th International Conference on Conceptual Structures, ICCS 2021, Virtual Event, September 20-22, 2021, Proceedings},

editor = {Braun, Tanya and Gehrke, Marcel and Hanika, Tom and Hernandez, Nathalie},

keywords = {2021 itegpub kde kdepub myown sai scale-measure scaling},

pages = {105--118},

publisher = {Springer},

series = {Lecture Notes in Computer Science},

title = {Quantifying the Conceptual Error in Dimensionality Reduction},

volume = 12879,

year = 2021

}%0 Conference Paper

%1 DBLP:conf/iccs/HanikaH21

%A Hanika, Tom

%A Hirth, Johannes

%B Graph-Based Representation and Reasoning - 26th International Conference on Conceptual Structures, ICCS 2021, Virtual Event, September 20-22, 2021, Proceedings

%D 2021

%E Braun, Tanya

%E Gehrke, Marcel

%E Hanika, Tom

%E Hernandez, Nathalie

%I Springer

%P 105--118

%R 10.1007/978-3-030-86982-3_8

%T Quantifying the Conceptual Error in Dimensionality Reduction

%U https://doi.org/10.1007/978-3-030-86982-3_8

%V 12879 - 1.Koyda, M., Stumme, G.: Boolean Substructures in Formal Concept Analysis. ICFCA: International Conference on Formal Concept Analysis. pp. 38–53. Springer (2021).
@conference{koyda2021boolean,

author = {Koyda, Maren and Stumme, Gerd},

booktitle = {Formal Concept Analysis - 16th International Conference, ICFCA 2021, Strasbourg, France, June 29 – July 2, 2021, Proceedings},

journal = {ICFCA: International Conference on Formal Concept Analysis},

keywords = {2021 fca itegpub myown},

pages = {38-53},

publisher = {Springer},

series = {Lecture Notes in Computer Science},

title = {Boolean Substructures in Formal Concept Analysis},

year = 2021

}%0 Generic

%1 koyda2021boolean

%A Koyda, Maren

%A Stumme, Gerd

%B Formal Concept Analysis - 16th International Conference, ICFCA 2021, Strasbourg, France, June 29 – July 2, 2021, Proceedings

%D 2021

%I Springer

%J ICFCA: International Conference on Formal Concept Analysis

%P 38-53

%T Boolean Substructures in Formal Concept Analysis

%@ 978-3-030-77866-8 - 1.Schäfermeier, B., Stumme, G., Hanika, T.: Towards Explainable Scientific Venue Recommendations, http://arxiv.org/abs/2109.11343, (2021).Selecting the best scientific venue (i.e., conference/journal) for the submission of a research article constitutes a multifaceted challenge. Important aspects to consider are the suitability of research topics, a venue's prestige, and the probability of acceptance. The selection problem is exacerbated through the continuous emergence of additional venues. Previously proposed approaches for supporting authors in this process rely on complex recommender systems, e.g., based on Word2Vec or TextCNN. These, however, often elude an explanation for their recommendations. In this work, we propose an unsophisticated method that advances the state-of-the-art in two aspects: First, we enhance the interpretability of recommendations through non-negative matrix factorization based topic models; Second, we surprisingly can obtain competitive recommendation performance while using simpler learning methods.
@misc{schafermeier2021towards,

abstract = {Selecting the best scientific venue (i.e., conference/journal) for the submission of a research article constitutes a multifaceted challenge. Important aspects to consider are the suitability of research topics, a venue's prestige, and the probability of acceptance. The selection problem is exacerbated through the continuous emergence of additional venues. Previously proposed approaches for supporting authors in this process rely on complex recommender systems, e.g., based on Word2Vec or TextCNN. These, however, often elude an explanation for their recommendations. In this work, we propose an unsophisticated method that advances the state-of-the-art in two aspects: First, we enhance the interpretability of recommendations through non-negative matrix factorization based topic models; Second, we surprisingly can obtain competitive recommendation performance while using simpler learning methods.},

author = {Schäfermeier, Bastian and Stumme, Gerd and Hanika, Tom},

keywords = {myown venue_recommendations},

note = {cite arxiv:2109.11343},

title = {Towards Explainable Scientific Venue Recommendations},

year = 2021

}%0 Generic

%1 schafermeier2021towards

%A Schäfermeier, Bastian

%A Stumme, Gerd

%A Hanika, Tom

%D 2021

%T Towards Explainable Scientific Venue Recommendations

%U http://arxiv.org/abs/2109.11343

%X Selecting the best scientific venue (i.e., conference/journal) for the submission of a research article constitutes a multifaceted challenge. Important aspects to consider are the suitability of research topics, a venue's prestige, and the probability of acceptance. The selection problem is exacerbated through the continuous emergence of additional venues. Previously proposed approaches for supporting authors in this process rely on complex recommender systems, e.g., based on Word2Vec or TextCNN. These, however, often elude an explanation for their recommendations. In this work, we propose an unsophisticated method that advances the state-of-the-art in two aspects: First, we enhance the interpretability of recommendations through non-negative matrix factorization based topic models; Second, we surprisingly can obtain competitive recommendation performance while using simpler learning methods. - 1.Dürrschnabel, D., Koyda, M., Stumme, G.: Attribute Selection Using Contranominal Scales. In: Braun, T., Gehrke, M., Hanika, T., and Hernandez, N. (eds.) Graph-Based Representation and Reasoning. pp. 127–141. Springer International Publishing, Cham (2021).Formal Concept Analysis (FCA) allows to analyze binary data by deriving concepts and ordering them in lattices. One of the main goals of FCA is to enable humans to comprehend the information that is encapsulated in the data; however, the large size of concept lattices is a limiting factor for the feasibility of understanding the underlying structural properties. The size of such a lattice depends on the number of subcontexts in the corresponding formal context that are isomorphic to a contranominal scale of high dimension. In this work, we propose the algorithm ContraFinder that enables the computation of all contranominal scales of a given formal context. Leveraging this algorithm, we introduce \\($\$\)backslashdelta \\($\$\)\($\delta$\)-adjusting, a novel approach in order to decrease the number of contranominal scales in a formal context by the selection of an appropriate attribute subset. We demonstrate that \\($\$\)backslashdelta \\($\$\)\($\delta$\)-adjusting a context reduces the size of the hereby emerging sub-semilattice and that the implication set is restricted to meaningful implications. This is evaluated with respect to its associated knowledge by means of a classification task. Hence, our proposed technique strongly improves understandability while preserving important conceptual structures.
@inproceedings{10.1007/978-3-030-86982-3_10,

abstract = {Formal Concept Analysis (FCA) allows to analyze binary data by deriving concepts and ordering them in lattices. One of the main goals of FCA is to enable humans to comprehend the information that is encapsulated in the data; however, the large size of concept lattices is a limiting factor for the feasibility of understanding the underlying structural properties. The size of such a lattice depends on the number of subcontexts in the corresponding formal context that are isomorphic to a contranominal scale of high dimension. In this work, we propose the algorithm ContraFinder that enables the computation of all contranominal scales of a given formal context. Leveraging this algorithm, we introduce \\($\$\)backslashdelta \\($\$\)\($\delta$\)-adjusting, a novel approach in order to decrease the number of contranominal scales in a formal context by the selection of an appropriate attribute subset. We demonstrate that \\($\$\)backslashdelta \\($\$\)\($\delta$\)-adjusting a context reduces the size of the hereby emerging sub-semilattice and that the implication set is restricted to meaningful implications. This is evaluated with respect to its associated knowledge by means of a classification task. Hence, our proposed technique strongly improves understandability while preserving important conceptual structures.},

address = {Cham},

author = {Dürrschnabel, Dominik and Koyda, Maren and Stumme, Gerd},

booktitle = {Graph-Based Representation and Reasoning},

editor = {Braun, Tanya and Gehrke, Marcel and Hanika, Tom and Hernandez, Nathalie},

keywords = {2021 fca kde myown},

pages = {127--141},

publisher = {Springer International Publishing},

title = {Attribute Selection Using Contranominal Scales},

year = 2021

}%0 Conference Paper

%1 10.1007/978-3-030-86982-3_10

%A Dürrschnabel, Dominik

%A Koyda, Maren

%A Stumme, Gerd

%B Graph-Based Representation and Reasoning

%C Cham

%D 2021

%E Braun, Tanya

%E Gehrke, Marcel

%E Hanika, Tom

%E Hernandez, Nathalie

%I Springer International Publishing

%P 127--141

%T Attribute Selection Using Contranominal Scales

%X Formal Concept Analysis (FCA) allows to analyze binary data by deriving concepts and ordering them in lattices. One of the main goals of FCA is to enable humans to comprehend the information that is encapsulated in the data; however, the large size of concept lattices is a limiting factor for the feasibility of understanding the underlying structural properties. The size of such a lattice depends on the number of subcontexts in the corresponding formal context that are isomorphic to a contranominal scale of high dimension. In this work, we propose the algorithm ContraFinder that enables the computation of all contranominal scales of a given formal context. Leveraging this algorithm, we introduce \\($\$\)backslashdelta \\($\$\)\($\delta$\)-adjusting, a novel approach in order to decrease the number of contranominal scales in a formal context by the selection of an appropriate attribute subset. We demonstrate that \\($\$\)backslashdelta \\($\$\)\($\delta$\)-adjusting a context reduces the size of the hereby emerging sub-semilattice and that the implication set is restricted to meaningful implications. This is evaluated with respect to its associated knowledge by means of a classification task. Hence, our proposed technique strongly improves understandability while preserving important conceptual structures.

%@ 978-3-030-86982-3 - 1.Draude, C., Gruhl, C., Hornung, G., Kropf, J., Lamla, J., Leimeister, J.M., Sick, B., Stumme, G.: Social Machines. Informatik Spektrum. (2021).
@article{2021,

author = {Draude, Claude and Gruhl, Christian and Hornung, Gerrit and Kropf, Jonathan and Lamla, Jörn and Leimeister, Jan Marco and Sick, Bernhard and Stumme, Gerd},

journal = {Informatik Spektrum},

keywords = {2021 itegpub myown},

month = {nov},

title = {Social Machines},

year = 2021

}%0 Journal Article

%1 2021

%A Draude, Claude

%A Gruhl, Christian

%A Hornung, Gerrit

%A Kropf, Jonathan

%A Lamla, Jörn

%A Leimeister, Jan Marco

%A Sick, Bernhard

%A Stumme, Gerd

%D 2021

%J Informatik Spektrum

%R 10.1007/s00287-021-01421-4

%T Social Machines

%U https://doi.org/10.1007%2Fs00287-021-01421-4 - 1.Koopmann, T., Stubbemann, M., Kapa, M., Paris, M., Buenstorf, G., Hanika, T., Hotho, A., Jäschke, R., Stumme, G.: Proximity dimensions and the emergence of collaboration: a HypTrails study on German AI research. Scientometrics. (2021).Creation and exchange of knowledge depends on collaboration. Recent work has suggested that the emergence of collaboration frequently relies on geographic proximity. However, being co-located tends to be associated with other dimensions of proximity, such as social ties or a shared organizational environment. To account for such factors, multiple dimensions of proximity have been proposed, including cognitive, institutional, organizational, social and geographical proximity. Since they strongly interrelate, disentangling these dimensions and their respective impact on collaboration is challenging. To address this issue, we propose various methods for measuring different dimensions of proximity. We then present an approach to compare and rank them with respect to the extent to which they indicate co-publications and co-inventions. We adapt the HypTrails approach, which was originally developed to explain human navigation, to co-author and co-inventor graphs. We evaluate this approach on a subset of the German research community, specifically academic authors and inventors active in research on artificial intelligence (AI). We find that social proximity and cognitive proximity are more important for the emergence of collaboration than geographic proximity.
@article{koopmann2021proximity,

abstract = {Creation and exchange of knowledge depends on collaboration. Recent work has suggested that the emergence of collaboration frequently relies on geographic proximity. However, being co-located tends to be associated with other dimensions of proximity, such as social ties or a shared organizational environment. To account for such factors, multiple dimensions of proximity have been proposed, including cognitive, institutional, organizational, social and geographical proximity. Since they strongly interrelate, disentangling these dimensions and their respective impact on collaboration is challenging. To address this issue, we propose various methods for measuring different dimensions of proximity. We then present an approach to compare and rank them with respect to the extent to which they indicate co-publications and co-inventions. We adapt the HypTrails approach, which was originally developed to explain human navigation, to co-author and co-inventor graphs. We evaluate this approach on a subset of the German research community, specifically academic authors and inventors active in research on artificial intelligence (AI). We find that social proximity and cognitive proximity are more important for the emergence of collaboration than geographic proximity.},

author = {Koopmann, Tobias and Stubbemann, Maximilian and Kapa, Matthias and Paris, Michael and Buenstorf, Guido and Hanika, Tom and Hotho, Andreas and Jäschke, Robert and Stumme, Gerd},

journal = {Scientometrics},

keywords = {2020 homepage kdepub myown regio},

title = {Proximity dimensions and the emergence of collaboration: a HypTrails study on German AI research},

year = 2021

}%0 Journal Article

%1 koopmann2021proximity

%A Koopmann, Tobias

%A Stubbemann, Maximilian

%A Kapa, Matthias

%A Paris, Michael

%A Buenstorf, Guido

%A Hanika, Tom

%A Hotho, Andreas

%A Jäschke, Robert

%A Stumme, Gerd

%D 2021

%J Scientometrics

%R 10.1007/s11192-021-03922-1

%T Proximity dimensions and the emergence of collaboration: a HypTrails study on German AI research

%U https://doi.org/10.1007/s11192-021-03922-1

%X Creation and exchange of knowledge depends on collaboration. Recent work has suggested that the emergence of collaboration frequently relies on geographic proximity. However, being co-located tends to be associated with other dimensions of proximity, such as social ties or a shared organizational environment. To account for such factors, multiple dimensions of proximity have been proposed, including cognitive, institutional, organizational, social and geographical proximity. Since they strongly interrelate, disentangling these dimensions and their respective impact on collaboration is challenging. To address this issue, we propose various methods for measuring different dimensions of proximity. We then present an approach to compare and rank them with respect to the extent to which they indicate co-publications and co-inventions. We adapt the HypTrails approach, which was originally developed to explain human navigation, to co-author and co-inventor graphs. We evaluate this approach on a subset of the German research community, specifically academic authors and inventors active in research on artificial intelligence (AI). We find that social proximity and cognitive proximity are more important for the emergence of collaboration than geographic proximity. - 1.Schaefermeier, B., Stumme, G., Hanika, T.: Topological Indoor Mapping through WiFi Signals. (2021).The ubiquitous presence of WiFi access points and mobile devices capable of measuring WiFi signal strengths allow for real-world applications in indoor localization and mapping. In particular, no additional infrastructure is required. Previous approaches in this field were, however, often hindered by problems such as effortful map-building processes, changing environments and hardware differences. We tackle these problems focussing on topological maps. These represent discrete locations, such as rooms, and their relations, e.g., distances and transition frequencies. In our unsupervised method, we employ WiFi signal strength distributions, dimension reduction and clustering. It can be used in settings where users carry mobile devices and follow their normal routine. We aim for applications in short-lived indoor events such as conferences.
@article{schaefermeier2021topological,

abstract = {The ubiquitous presence of WiFi access points and mobile devices capable of measuring WiFi signal strengths allow for real-world applications in indoor localization and mapping. In particular, no additional infrastructure is required. Previous approaches in this field were, however, often hindered by problems such as effortful map-building processes, changing environments and hardware differences. We tackle these problems focussing on topological maps. These represent discrete locations, such as rooms, and their relations, e.g., distances and transition frequencies. In our unsupervised method, we employ WiFi signal strength distributions, dimension reduction and clustering. It can be used in settings where users carry mobile devices and follow their normal routine. We aim for applications in short-lived indoor events such as conferences.},

author = {Schaefermeier, Bastian and Stumme, Gerd and Hanika, Tom},

keywords = {mapping myown wifi},

note = {cite arxiv:2106.09789Comment: 18 pages},

title = {Topological Indoor Mapping through WiFi Signals},

year = 2021

}%0 Journal Article

%1 schaefermeier2021topological

%A Schaefermeier, Bastian

%A Stumme, Gerd

%A Hanika, Tom

%D 2021

%T Topological Indoor Mapping through WiFi Signals

%U http://arxiv.org/abs/2106.09789

%X The ubiquitous presence of WiFi access points and mobile devices capable of measuring WiFi signal strengths allow for real-world applications in indoor localization and mapping. In particular, no additional infrastructure is required. Previous approaches in this field were, however, often hindered by problems such as effortful map-building processes, changing environments and hardware differences. We tackle these problems focussing on topological maps. These represent discrete locations, such as rooms, and their relations, e.g., distances and transition frequencies. In our unsupervised method, we employ WiFi signal strength distributions, dimension reduction and clustering. It can be used in settings where users carry mobile devices and follow their normal routine. We aim for applications in short-lived indoor events such as conferences. - 1.Dürrschnabel, D., Hanika, T., Stubbemann, M.: FCA2VEC: Embedding Techniques for Formal Concept Analysis. Presented at the (2021).
@inbook{durrschnabel2021fca2vec,

author = {Dürrschnabel, Dominik and Hanika, Tom and Stubbemann, Maximilian},

keywords = {2021 closure_operator fca2vec myown word2vec},

title = {FCA2VEC: Embedding Techniques for Formal Concept Analysis},

year = 2021

}%0 Book Section

%1 durrschnabel2021fca2vec

%A Dürrschnabel, Dominik

%A Hanika, Tom

%A Stubbemann, Maximilian

%D 2021

%T FCA2VEC: Embedding Techniques for Formal Concept Analysis