Portfolio item number 1
Short description of portfolio item number 1
Short description of portfolio item number 1
Short description of portfolio item number 2
Published in IEEE Transactions on Pattern Analysis and Machine Intelligence (under review), 2023
How to implement partial equivariant model for disentanglement learning?
Published in Expert Systems with Applications, 2023
How to transfer the feature structure from teacher to student model?
BibTeX: @article{JUNG2023120980, title = {Feature structure distillation with Centered Kernel Alignment in BERT transferring}, journal = {Expert Systems with Applications}, volume = {234}, pages = {120980}, year = {2023}, issn = {0957-4174}, doi = {https://doi.org/10.1016/j.eswa.2023.120980}, url = {https://www.sciencedirect.com/science/article/pii/S0957417423014823}, author = {Hee-Jun Jung and Doyeon Kim and Seung-Hoon Na and Kangil Kim}, keywords = {Knowledge distillation, BERT, Centered Kernel Alignment, Natural language processing}, abstract = {Knowledge distillation is an approach to transfer information on representations from a teacher to a student by reducing their difference. A challenge of this approach is to reduce the flexibility of the student’s representations inducing inaccurate learning of the teacher’s knowledge. To resolve the problems, we propose a novel method feature structure distillation that elaborates information on structures of features into three types for transferring, and implements them based on Centered Kernel Analysis. In particular, the global local-inter structure is proposed to transfer the structure beyond the mini-batch. In detail, the method first divides the feature information into three structures: intra-feature, local inter-feature, and global inter-feature structures to subdivide the structure and transfer the diversity of the structure. Then, we adopt CKA which shows a more accurate similarity metric compared to other metrics between two different models or representations on different spaces. In particular, a memory-augmented transfer method with clustering is implemented for the global structures. The methods are empirically analyzed on the nine tasks for language understanding of the GLUE dataset with Bidirectional Encoder Representations from Transformers (BERT), which is a representative neural language model. In the results, the proposed methods effectively transfer the three types of structures and improves performance compared to state-of-the-art distillation methods: (i.e.) ours achieve 66.61% accuracy compared to the baseline (65.55%) in the RTE dataset. Indeed, the code for the methods is available at https://github.com/maroo-sky/FSD.} }
Paper | Code
Published in TMLR, 2024
How to represent symmetries for disentanglement learning without ground truth?
BibTeX: @article{ jung2024cfasl, title={{CFASL}: Composite Factor-Aligned Symmetry Learning for Disentanglement in Variational AutoEncoder}, author={Hee-Jun Jung and Jaehyoung Jeong and Kangil Kim}, journal={Transactions on Machine Learning Research}, issn={2835-8856}, year={2024}, url={https://openreview.net/forum?id=mDGvrH7lju}, note={} }
Paper | Code
Published:
This is a description of your talk, which is a markdown files that can be all markdown-ified like any other post. Yay markdown!
Published:
This is a description of your conference proceedings talk, note the different field in type. You can put anything in this field.
Graduate Course, GIST, AIGS, 2022
This cource gives comprehensive lectures and projects about NLP. (All materials have been kept strictly confidential.)
03/03 Lecture 1. Course Introduction
03/03 Lecture 2. Programming Language Basics for Project
03/08 Lecture 3. Overview of AI
03/10-03/15 Lecture 4. Problem Solving
03/17 Lecture 5. Model Representation
03/22-03/31 Lecture 6. Optimization & Generalization
04/05 Lecture 7. Evaluation
04/07-04/14 Lecture 8. Probabilistic Modeling
Basics
Bayesian Network
Markov Random Field
04/19-04/21 Lecture 9. Logic Modeling
Background
Predicate Calculus
Inductive Logit Programming
Mid-term exam -> Replaced to Lab 2 Project
04/26-04/28 Lecture 10. Overview of SRL
Hyperplane-based Modeling - Neural Networks
Why SRL?
05/03-05/12 Lecture 11. Integration of Logic and Probability
Bayesian Logic Programming
Markov Logic Network
Logic Modeling: Basics
Logic Modeling: Formal Grammar
Probabilistic Context Free Grammar
05/17-05/26 Lecture 12. Overview of Natural Language Processing
Linguistics Essentials + More Terms
NLP Problems and Their Relations
N-gram
05/31 Lecture 13. Language Model
06/02 Lecture 14. Machine Translation
06/07 Lecture 15. Question Answering
06/09 Lecture 16. Conversational System
06/14 Final-Term Exam
Project contents introduce basic practical knowledge different to the contents of the lectures. To cover the simple knowledge and programming issues, we will have 6 lab times in this semester (in our class time, no additional lab times are assigned.)
CHECK the key date of each assignment: (To be uploaded; Q&A Session; Deadline)